00:00:00.000 Started by upstream project "autotest-per-patch" build number 130589 00:00:00.000 originally caused by: 00:00:00.001 Started by user sys_sgci 00:00:00.023 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-vg.groovy 00:00:14.157 The recommended git tool is: git 00:00:14.157 using credential 00000000-0000-0000-0000-000000000002 00:00:14.159 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:14.169 Fetching changes from the remote Git repository 00:00:14.173 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:14.184 Using shallow fetch with depth 1 00:00:14.184 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:14.184 > git --version # timeout=10 00:00:14.194 > git --version # 'git version 2.39.2' 00:00:14.194 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:14.207 Setting http proxy: proxy-dmz.intel.com:911 00:00:14.207 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:28.906 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:28.919 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:28.931 Checking out Revision d37d6e8a0abef39b377a5f0531b43b2efbbebf34 (FETCH_HEAD) 00:00:28.932 > git config core.sparsecheckout # timeout=10 00:00:28.943 > git read-tree -mu HEAD # timeout=10 00:00:28.960 > git checkout -f d37d6e8a0abef39b377a5f0531b43b2efbbebf34 # timeout=5 00:00:28.983 Commit message: "pool: serialize build page context to json" 00:00:28.983 > git rev-list --no-walk d37d6e8a0abef39b377a5f0531b43b2efbbebf34 # timeout=10 00:00:29.091 [Pipeline] Start of Pipeline 00:00:29.101 [Pipeline] library 00:00:29.102 Loading library shm_lib@master 00:00:29.102 Library shm_lib@master is cached. Copying from home. 00:00:29.116 [Pipeline] node 00:00:44.117 Still waiting to schedule task 00:00:44.118 Waiting for next available executor on ‘vagrant-vm-host’ 00:12:18.890 Running on VM-host-SM17 in /var/jenkins/workspace/raid-vg-autotest_3 00:12:18.891 [Pipeline] { 00:12:18.898 [Pipeline] catchError 00:12:18.899 [Pipeline] { 00:12:18.910 [Pipeline] wrap 00:12:18.921 [Pipeline] { 00:12:18.931 [Pipeline] stage 00:12:18.933 [Pipeline] { (Prologue) 00:12:18.951 [Pipeline] echo 00:12:18.953 Node: VM-host-SM17 00:12:18.959 [Pipeline] cleanWs 00:12:18.968 [WS-CLEANUP] Deleting project workspace... 00:12:18.968 [WS-CLEANUP] Deferred wipeout is used... 00:12:18.994 [WS-CLEANUP] done 00:12:19.200 [Pipeline] setCustomBuildProperty 00:12:19.263 [Pipeline] httpRequest 00:12:19.661 [Pipeline] echo 00:12:19.663 Sorcerer 10.211.164.101 is alive 00:12:19.674 [Pipeline] retry 00:12:19.676 [Pipeline] { 00:12:19.690 [Pipeline] httpRequest 00:12:19.695 HttpMethod: GET 00:12:19.695 URL: http://10.211.164.101/packages/jbp_d37d6e8a0abef39b377a5f0531b43b2efbbebf34.tar.gz 00:12:19.696 Sending request to url: http://10.211.164.101/packages/jbp_d37d6e8a0abef39b377a5f0531b43b2efbbebf34.tar.gz 00:12:19.697 Response Code: HTTP/1.1 200 OK 00:12:19.698 Success: Status code 200 is in the accepted range: 200,404 00:12:19.699 Saving response body to /var/jenkins/workspace/raid-vg-autotest_3/jbp_d37d6e8a0abef39b377a5f0531b43b2efbbebf34.tar.gz 00:12:19.986 [Pipeline] } 00:12:20.004 [Pipeline] // retry 00:12:20.012 [Pipeline] sh 00:12:20.292 + tar --no-same-owner -xf jbp_d37d6e8a0abef39b377a5f0531b43b2efbbebf34.tar.gz 00:12:20.308 [Pipeline] httpRequest 00:12:20.709 [Pipeline] echo 00:12:20.711 Sorcerer 10.211.164.101 is alive 00:12:20.721 [Pipeline] retry 00:12:20.723 [Pipeline] { 00:12:20.738 [Pipeline] httpRequest 00:12:20.743 HttpMethod: GET 00:12:20.744 URL: http://10.211.164.101/packages/spdk_1b1c3081e7433ef3ee5ea712b81b554bbbca8f0a.tar.gz 00:12:20.745 Sending request to url: http://10.211.164.101/packages/spdk_1b1c3081e7433ef3ee5ea712b81b554bbbca8f0a.tar.gz 00:12:20.745 Response Code: HTTP/1.1 200 OK 00:12:20.746 Success: Status code 200 is in the accepted range: 200,404 00:12:20.747 Saving response body to /var/jenkins/workspace/raid-vg-autotest_3/spdk_1b1c3081e7433ef3ee5ea712b81b554bbbca8f0a.tar.gz 00:12:23.318 [Pipeline] } 00:12:23.333 [Pipeline] // retry 00:12:23.340 [Pipeline] sh 00:12:23.614 + tar --no-same-owner -xf spdk_1b1c3081e7433ef3ee5ea712b81b554bbbca8f0a.tar.gz 00:12:26.948 [Pipeline] sh 00:12:27.242 + git -C spdk log --oneline -n5 00:12:27.242 1b1c3081e bdev: explicitly inline bdev_channel_get_io() 00:12:27.242 165425556 bdev/passthru: add bdev_io_stack support 00:12:27.242 4f975f22c [TEST] bdev: save stack_frame instead of bdev_io in io_submitted TAILQ 00:12:27.242 392076696 bdev: Add spdk_bdev_io_submit API 00:12:27.242 ef2413376 bdev: Add spdk_bdev_io_to_ctx 00:12:27.259 [Pipeline] writeFile 00:12:27.274 [Pipeline] sh 00:12:27.553 + jbp/jenkins/jjb-config/jobs/scripts/autorun_quirks.sh 00:12:27.564 [Pipeline] sh 00:12:27.844 + cat autorun-spdk.conf 00:12:27.844 SPDK_RUN_FUNCTIONAL_TEST=1 00:12:27.844 SPDK_RUN_ASAN=1 00:12:27.844 SPDK_RUN_UBSAN=1 00:12:27.844 SPDK_TEST_RAID=1 00:12:27.844 SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:12:27.850 RUN_NIGHTLY=0 00:12:27.852 [Pipeline] } 00:12:27.869 [Pipeline] // stage 00:12:27.884 [Pipeline] stage 00:12:27.886 [Pipeline] { (Run VM) 00:12:27.899 [Pipeline] sh 00:12:28.224 + jbp/jenkins/jjb-config/jobs/scripts/prepare_nvme.sh 00:12:28.224 + echo 'Start stage prepare_nvme.sh' 00:12:28.224 Start stage prepare_nvme.sh 00:12:28.224 + [[ -n 6 ]] 00:12:28.224 + disk_prefix=ex6 00:12:28.224 + [[ -n /var/jenkins/workspace/raid-vg-autotest_3 ]] 00:12:28.224 + [[ -e /var/jenkins/workspace/raid-vg-autotest_3/autorun-spdk.conf ]] 00:12:28.224 + source /var/jenkins/workspace/raid-vg-autotest_3/autorun-spdk.conf 00:12:28.224 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:12:28.224 ++ SPDK_RUN_ASAN=1 00:12:28.224 ++ SPDK_RUN_UBSAN=1 00:12:28.224 ++ SPDK_TEST_RAID=1 00:12:28.224 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:12:28.224 ++ RUN_NIGHTLY=0 00:12:28.224 + cd /var/jenkins/workspace/raid-vg-autotest_3 00:12:28.224 + nvme_files=() 00:12:28.224 + declare -A nvme_files 00:12:28.224 + backend_dir=/var/lib/libvirt/images/backends 00:12:28.224 + nvme_files['nvme.img']=5G 00:12:28.224 + nvme_files['nvme-cmb.img']=5G 00:12:28.224 + nvme_files['nvme-multi0.img']=4G 00:12:28.224 + nvme_files['nvme-multi1.img']=4G 00:12:28.224 + nvme_files['nvme-multi2.img']=4G 00:12:28.224 + nvme_files['nvme-openstack.img']=8G 00:12:28.224 + nvme_files['nvme-zns.img']=5G 00:12:28.224 + (( SPDK_TEST_NVME_PMR == 1 )) 00:12:28.224 + (( SPDK_TEST_FTL == 1 )) 00:12:28.224 + (( SPDK_TEST_NVME_FDP == 1 )) 00:12:28.224 + [[ ! -d /var/lib/libvirt/images/backends ]] 00:12:28.224 + for nvme in "${!nvme_files[@]}" 00:12:28.224 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-multi2.img -s 4G 00:12:28.224 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-multi2.img', fmt=raw size=4294967296 preallocation=falloc 00:12:28.224 + for nvme in "${!nvme_files[@]}" 00:12:28.224 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-cmb.img -s 5G 00:12:28.224 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-cmb.img', fmt=raw size=5368709120 preallocation=falloc 00:12:28.224 + for nvme in "${!nvme_files[@]}" 00:12:28.224 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-openstack.img -s 8G 00:12:28.224 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-openstack.img', fmt=raw size=8589934592 preallocation=falloc 00:12:28.224 + for nvme in "${!nvme_files[@]}" 00:12:28.224 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-zns.img -s 5G 00:12:28.224 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-zns.img', fmt=raw size=5368709120 preallocation=falloc 00:12:28.224 + for nvme in "${!nvme_files[@]}" 00:12:28.224 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-multi1.img -s 4G 00:12:28.224 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-multi1.img', fmt=raw size=4294967296 preallocation=falloc 00:12:28.224 + for nvme in "${!nvme_files[@]}" 00:12:28.224 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-multi0.img -s 4G 00:12:28.224 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-multi0.img', fmt=raw size=4294967296 preallocation=falloc 00:12:28.224 + for nvme in "${!nvme_files[@]}" 00:12:28.224 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme.img -s 5G 00:12:28.224 Formatting '/var/lib/libvirt/images/backends/ex6-nvme.img', fmt=raw size=5368709120 preallocation=falloc 00:12:28.224 ++ sudo grep -rl ex6-nvme.img /etc/libvirt/qemu 00:12:28.224 + echo 'End stage prepare_nvme.sh' 00:12:28.224 End stage prepare_nvme.sh 00:12:28.236 [Pipeline] sh 00:12:28.514 + DISTRO=fedora39 CPUS=10 RAM=12288 jbp/jenkins/jjb-config/jobs/scripts/vagrant_create_vm.sh 00:12:28.514 Setup: -n 10 -s 12288 -x http://proxy-dmz.intel.com:911 -p libvirt --qemu-emulator=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 --nic-model=e1000 -b /var/lib/libvirt/images/backends/ex6-nvme.img -b /var/lib/libvirt/images/backends/ex6-nvme-multi0.img,nvme,/var/lib/libvirt/images/backends/ex6-nvme-multi1.img:/var/lib/libvirt/images/backends/ex6-nvme-multi2.img -H -a -v -f fedora39 00:12:28.514 00:12:28.514 DIR=/var/jenkins/workspace/raid-vg-autotest_3/spdk/scripts/vagrant 00:12:28.514 SPDK_DIR=/var/jenkins/workspace/raid-vg-autotest_3/spdk 00:12:28.514 VAGRANT_TARGET=/var/jenkins/workspace/raid-vg-autotest_3 00:12:28.514 HELP=0 00:12:28.514 DRY_RUN=0 00:12:28.514 NVME_FILE=/var/lib/libvirt/images/backends/ex6-nvme.img,/var/lib/libvirt/images/backends/ex6-nvme-multi0.img, 00:12:28.514 NVME_DISKS_TYPE=nvme,nvme, 00:12:28.514 NVME_AUTO_CREATE=0 00:12:28.514 NVME_DISKS_NAMESPACES=,/var/lib/libvirt/images/backends/ex6-nvme-multi1.img:/var/lib/libvirt/images/backends/ex6-nvme-multi2.img, 00:12:28.514 NVME_CMB=,, 00:12:28.514 NVME_PMR=,, 00:12:28.514 NVME_ZNS=,, 00:12:28.514 NVME_MS=,, 00:12:28.514 NVME_FDP=,, 00:12:28.514 SPDK_VAGRANT_DISTRO=fedora39 00:12:28.514 SPDK_VAGRANT_VMCPU=10 00:12:28.514 SPDK_VAGRANT_VMRAM=12288 00:12:28.514 SPDK_VAGRANT_PROVIDER=libvirt 00:12:28.514 SPDK_VAGRANT_HTTP_PROXY=http://proxy-dmz.intel.com:911 00:12:28.514 SPDK_QEMU_EMULATOR=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 00:12:28.514 SPDK_OPENSTACK_NETWORK=0 00:12:28.514 VAGRANT_PACKAGE_BOX=0 00:12:28.514 VAGRANTFILE=/var/jenkins/workspace/raid-vg-autotest_3/spdk/scripts/vagrant/Vagrantfile 00:12:28.514 FORCE_DISTRO=true 00:12:28.514 VAGRANT_BOX_VERSION= 00:12:28.514 EXTRA_VAGRANTFILES= 00:12:28.514 NIC_MODEL=e1000 00:12:28.514 00:12:28.514 mkdir: created directory '/var/jenkins/workspace/raid-vg-autotest_3/fedora39-libvirt' 00:12:28.514 /var/jenkins/workspace/raid-vg-autotest_3/fedora39-libvirt /var/jenkins/workspace/raid-vg-autotest_3 00:12:31.799 Bringing machine 'default' up with 'libvirt' provider... 00:12:32.455 ==> default: Creating image (snapshot of base box volume). 00:12:32.714 ==> default: Creating domain with the following settings... 00:12:32.714 ==> default: -- Name: fedora39-39-1.5-1721788873-2326_default_1727813487_4fa61dc8df2066f22b57 00:12:32.714 ==> default: -- Domain type: kvm 00:12:32.714 ==> default: -- Cpus: 10 00:12:32.714 ==> default: -- Feature: acpi 00:12:32.714 ==> default: -- Feature: apic 00:12:32.714 ==> default: -- Feature: pae 00:12:32.714 ==> default: -- Memory: 12288M 00:12:32.714 ==> default: -- Memory Backing: hugepages: 00:12:32.714 ==> default: -- Management MAC: 00:12:32.714 ==> default: -- Loader: 00:12:32.714 ==> default: -- Nvram: 00:12:32.714 ==> default: -- Base box: spdk/fedora39 00:12:32.714 ==> default: -- Storage pool: default 00:12:32.714 ==> default: -- Image: /var/lib/libvirt/images/fedora39-39-1.5-1721788873-2326_default_1727813487_4fa61dc8df2066f22b57.img (20G) 00:12:32.714 ==> default: -- Volume Cache: default 00:12:32.714 ==> default: -- Kernel: 00:12:32.714 ==> default: -- Initrd: 00:12:32.714 ==> default: -- Graphics Type: vnc 00:12:32.714 ==> default: -- Graphics Port: -1 00:12:32.714 ==> default: -- Graphics IP: 127.0.0.1 00:12:32.714 ==> default: -- Graphics Password: Not defined 00:12:32.714 ==> default: -- Video Type: cirrus 00:12:32.714 ==> default: -- Video VRAM: 9216 00:12:32.714 ==> default: -- Sound Type: 00:12:32.714 ==> default: -- Keymap: en-us 00:12:32.714 ==> default: -- TPM Path: 00:12:32.714 ==> default: -- INPUT: type=mouse, bus=ps2 00:12:32.714 ==> default: -- Command line args: 00:12:32.714 ==> default: -> value=-device, 00:12:32.714 ==> default: -> value=nvme,id=nvme-0,serial=12340,addr=0x10, 00:12:32.714 ==> default: -> value=-drive, 00:12:32.714 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex6-nvme.img,if=none,id=nvme-0-drive0, 00:12:32.714 ==> default: -> value=-device, 00:12:32.714 ==> default: -> value=nvme-ns,drive=nvme-0-drive0,bus=nvme-0,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:12:32.714 ==> default: -> value=-device, 00:12:32.714 ==> default: -> value=nvme,id=nvme-1,serial=12341,addr=0x11, 00:12:32.714 ==> default: -> value=-drive, 00:12:32.714 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex6-nvme-multi0.img,if=none,id=nvme-1-drive0, 00:12:32.714 ==> default: -> value=-device, 00:12:32.714 ==> default: -> value=nvme-ns,drive=nvme-1-drive0,bus=nvme-1,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:12:32.714 ==> default: -> value=-drive, 00:12:32.714 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex6-nvme-multi1.img,if=none,id=nvme-1-drive1, 00:12:32.714 ==> default: -> value=-device, 00:12:32.714 ==> default: -> value=nvme-ns,drive=nvme-1-drive1,bus=nvme-1,nsid=2,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:12:32.714 ==> default: -> value=-drive, 00:12:32.714 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex6-nvme-multi2.img,if=none,id=nvme-1-drive2, 00:12:32.714 ==> default: -> value=-device, 00:12:32.714 ==> default: -> value=nvme-ns,drive=nvme-1-drive2,bus=nvme-1,nsid=3,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:12:32.973 ==> default: Creating shared folders metadata... 00:12:32.973 ==> default: Starting domain. 00:12:34.876 ==> default: Waiting for domain to get an IP address... 00:12:53.026 ==> default: Waiting for SSH to become available... 00:12:53.026 ==> default: Configuring and enabling network interfaces... 00:12:56.315 default: SSH address: 192.168.121.46:22 00:12:56.315 default: SSH username: vagrant 00:12:56.315 default: SSH auth method: private key 00:12:58.218 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest_3/spdk/ => /home/vagrant/spdk_repo/spdk 00:13:06.332 ==> default: Mounting SSHFS shared folder... 00:13:08.237 ==> default: Mounting folder via SSHFS: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest_3/fedora39-libvirt/output => /home/vagrant/spdk_repo/output 00:13:08.237 ==> default: Checking Mount.. 00:13:09.173 ==> default: Folder Successfully Mounted! 00:13:09.173 ==> default: Running provisioner: file... 00:13:10.109 default: ~/.gitconfig => .gitconfig 00:13:10.368 00:13:10.368 SUCCESS! 00:13:10.368 00:13:10.368 cd to /var/jenkins/workspace/raid-vg-autotest_3/fedora39-libvirt and type "vagrant ssh" to use. 00:13:10.368 Use vagrant "suspend" and vagrant "resume" to stop and start. 00:13:10.368 Use vagrant "destroy" followed by "rm -rf /var/jenkins/workspace/raid-vg-autotest_3/fedora39-libvirt" to destroy all trace of vm. 00:13:10.368 00:13:10.377 [Pipeline] } 00:13:10.396 [Pipeline] // stage 00:13:10.404 [Pipeline] dir 00:13:10.404 Running in /var/jenkins/workspace/raid-vg-autotest_3/fedora39-libvirt 00:13:10.406 [Pipeline] { 00:13:10.418 [Pipeline] catchError 00:13:10.420 [Pipeline] { 00:13:10.431 [Pipeline] sh 00:13:10.708 + vagrant ssh-config --host vagrant 00:13:10.708 + sed -ne /^Host/,$p 00:13:10.708 + tee ssh_conf 00:13:14.894 Host vagrant 00:13:14.894 HostName 192.168.121.46 00:13:14.894 User vagrant 00:13:14.894 Port 22 00:13:14.894 UserKnownHostsFile /dev/null 00:13:14.894 StrictHostKeyChecking no 00:13:14.894 PasswordAuthentication no 00:13:14.894 IdentityFile /var/lib/libvirt/images/.vagrant.d/boxes/spdk-VAGRANTSLASH-fedora39/39-1.5-1721788873-2326/libvirt/fedora39 00:13:14.894 IdentitiesOnly yes 00:13:14.894 LogLevel FATAL 00:13:14.894 ForwardAgent yes 00:13:14.894 ForwardX11 yes 00:13:14.894 00:13:14.906 [Pipeline] withEnv 00:13:14.908 [Pipeline] { 00:13:14.920 [Pipeline] sh 00:13:15.198 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant #!/bin/bash 00:13:15.198 source /etc/os-release 00:13:15.198 [[ -e /image.version ]] && img=$(< /image.version) 00:13:15.198 # Minimal, systemd-like check. 00:13:15.198 if [[ -e /.dockerenv ]]; then 00:13:15.198 # Clear garbage from the node's name: 00:13:15.198 # agt-er_autotest_547-896 -> autotest_547-896 00:13:15.198 # $HOSTNAME is the actual container id 00:13:15.198 agent=$HOSTNAME@${DOCKER_SWARM_PLUGIN_JENKINS_AGENT_NAME#*_} 00:13:15.198 if grep -q "/etc/hostname" /proc/self/mountinfo; then 00:13:15.198 # We can assume this is a mount from a host where container is running, 00:13:15.198 # so fetch its hostname to easily identify the target swarm worker. 00:13:15.198 container="$(< /etc/hostname) ($agent)" 00:13:15.198 else 00:13:15.198 # Fallback 00:13:15.198 container=$agent 00:13:15.198 fi 00:13:15.198 fi 00:13:15.198 echo "${NAME} ${VERSION_ID}|$(uname -r)|${img:-N/A}|${container:-N/A}" 00:13:15.198 00:13:15.463 [Pipeline] } 00:13:15.473 [Pipeline] // withEnv 00:13:15.478 [Pipeline] setCustomBuildProperty 00:13:15.489 [Pipeline] stage 00:13:15.491 [Pipeline] { (Tests) 00:13:15.505 [Pipeline] sh 00:13:15.782 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest_3/jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh vagrant@vagrant:./ 00:13:16.118 [Pipeline] sh 00:13:16.395 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest_3/jbp/jenkins/jjb-config/jobs/scripts/pkgdep-autoruner.sh vagrant@vagrant:./ 00:13:16.667 [Pipeline] timeout 00:13:16.667 Timeout set to expire in 1 hr 30 min 00:13:16.669 [Pipeline] { 00:13:16.682 [Pipeline] sh 00:13:16.961 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant git -C spdk_repo/spdk reset --hard 00:13:17.527 HEAD is now at 1b1c3081e bdev: explicitly inline bdev_channel_get_io() 00:13:17.620 [Pipeline] sh 00:13:17.897 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant sudo chown vagrant:vagrant spdk_repo 00:13:18.168 [Pipeline] sh 00:13:18.446 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest_3/autorun-spdk.conf vagrant@vagrant:spdk_repo 00:13:18.721 [Pipeline] sh 00:13:18.996 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant JOB_BASE_NAME=raid-vg-autotest ./autoruner.sh spdk_repo 00:13:19.256 ++ readlink -f spdk_repo 00:13:19.256 + DIR_ROOT=/home/vagrant/spdk_repo 00:13:19.256 + [[ -n /home/vagrant/spdk_repo ]] 00:13:19.256 + DIR_SPDK=/home/vagrant/spdk_repo/spdk 00:13:19.256 + DIR_OUTPUT=/home/vagrant/spdk_repo/output 00:13:19.256 + [[ -d /home/vagrant/spdk_repo/spdk ]] 00:13:19.256 + [[ ! -d /home/vagrant/spdk_repo/output ]] 00:13:19.256 + [[ -d /home/vagrant/spdk_repo/output ]] 00:13:19.256 + [[ raid-vg-autotest == pkgdep-* ]] 00:13:19.256 + cd /home/vagrant/spdk_repo 00:13:19.256 + source /etc/os-release 00:13:19.256 ++ NAME='Fedora Linux' 00:13:19.256 ++ VERSION='39 (Cloud Edition)' 00:13:19.256 ++ ID=fedora 00:13:19.256 ++ VERSION_ID=39 00:13:19.256 ++ VERSION_CODENAME= 00:13:19.256 ++ PLATFORM_ID=platform:f39 00:13:19.256 ++ PRETTY_NAME='Fedora Linux 39 (Cloud Edition)' 00:13:19.256 ++ ANSI_COLOR='0;38;2;60;110;180' 00:13:19.256 ++ LOGO=fedora-logo-icon 00:13:19.256 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:39 00:13:19.256 ++ HOME_URL=https://fedoraproject.org/ 00:13:19.256 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f39/system-administrators-guide/ 00:13:19.256 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:13:19.256 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:13:19.256 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:13:19.256 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=39 00:13:19.256 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:13:19.256 ++ REDHAT_SUPPORT_PRODUCT_VERSION=39 00:13:19.256 ++ SUPPORT_END=2024-11-12 00:13:19.256 ++ VARIANT='Cloud Edition' 00:13:19.256 ++ VARIANT_ID=cloud 00:13:19.256 + uname -a 00:13:19.256 Linux fedora39-cloud-1721788873-2326 6.8.9-200.fc39.x86_64 #1 SMP PREEMPT_DYNAMIC Wed Jul 24 03:04:40 UTC 2024 x86_64 GNU/Linux 00:13:19.256 + sudo /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:13:19.822 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:13:19.822 Hugepages 00:13:19.822 node hugesize free / total 00:13:19.822 node0 1048576kB 0 / 0 00:13:19.822 node0 2048kB 0 / 0 00:13:19.822 00:13:19.822 Type BDF Vendor Device NUMA Driver Device Block devices 00:13:19.822 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:13:19.822 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:13:19.822 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:13:19.822 + rm -f /tmp/spdk-ld-path 00:13:19.822 + source autorun-spdk.conf 00:13:19.822 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:13:19.822 ++ SPDK_RUN_ASAN=1 00:13:19.822 ++ SPDK_RUN_UBSAN=1 00:13:19.822 ++ SPDK_TEST_RAID=1 00:13:19.822 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:13:19.822 ++ RUN_NIGHTLY=0 00:13:19.822 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:13:19.822 + [[ -n '' ]] 00:13:19.822 + sudo git config --global --add safe.directory /home/vagrant/spdk_repo/spdk 00:13:19.822 + for M in /var/spdk/build-*-manifest.txt 00:13:19.822 + [[ -f /var/spdk/build-kernel-manifest.txt ]] 00:13:19.822 + cp /var/spdk/build-kernel-manifest.txt /home/vagrant/spdk_repo/output/ 00:13:19.822 + for M in /var/spdk/build-*-manifest.txt 00:13:19.822 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:13:19.822 + cp /var/spdk/build-pkg-manifest.txt /home/vagrant/spdk_repo/output/ 00:13:19.822 + for M in /var/spdk/build-*-manifest.txt 00:13:19.822 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:13:19.822 + cp /var/spdk/build-repo-manifest.txt /home/vagrant/spdk_repo/output/ 00:13:19.822 ++ uname 00:13:19.822 + [[ Linux == \L\i\n\u\x ]] 00:13:19.822 + sudo dmesg -T 00:13:19.822 + sudo dmesg --clear 00:13:19.822 + dmesg_pid=5207 00:13:19.822 + sudo dmesg -Tw 00:13:19.822 + [[ Fedora Linux == FreeBSD ]] 00:13:19.822 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:13:19.822 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:13:19.822 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:13:19.822 + [[ -x /usr/src/fio-static/fio ]] 00:13:19.822 + export FIO_BIN=/usr/src/fio-static/fio 00:13:19.822 + FIO_BIN=/usr/src/fio-static/fio 00:13:19.822 + [[ '' == \/\q\e\m\u\_\v\f\i\o\/* ]] 00:13:19.822 + [[ ! -v VFIO_QEMU_BIN ]] 00:13:19.822 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:13:19.822 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:13:19.822 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:13:19.822 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:13:19.822 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:13:19.822 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:13:19.822 + spdk/autorun.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:13:19.822 Test configuration: 00:13:19.822 SPDK_RUN_FUNCTIONAL_TEST=1 00:13:19.822 SPDK_RUN_ASAN=1 00:13:19.822 SPDK_RUN_UBSAN=1 00:13:19.822 SPDK_TEST_RAID=1 00:13:19.822 SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:13:20.081 RUN_NIGHTLY=0 20:12:15 -- common/autotest_common.sh@1680 -- $ [[ n == y ]] 00:13:20.081 20:12:15 -- common/autobuild_common.sh@15 -- $ source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:13:20.081 20:12:15 -- scripts/common.sh@15 -- $ shopt -s extglob 00:13:20.081 20:12:15 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:13:20.081 20:12:15 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:20.081 20:12:15 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:20.081 20:12:15 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:20.081 20:12:15 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:20.081 20:12:15 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:20.081 20:12:15 -- paths/export.sh@5 -- $ export PATH 00:13:20.081 20:12:15 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:20.081 20:12:15 -- common/autobuild_common.sh@478 -- $ out=/home/vagrant/spdk_repo/spdk/../output 00:13:20.081 20:12:15 -- common/autobuild_common.sh@479 -- $ date +%s 00:13:20.081 20:12:15 -- common/autobuild_common.sh@479 -- $ mktemp -dt spdk_1727813535.XXXXXX 00:13:20.081 20:12:15 -- common/autobuild_common.sh@479 -- $ SPDK_WORKSPACE=/tmp/spdk_1727813535.UMd48s 00:13:20.081 20:12:15 -- common/autobuild_common.sh@481 -- $ [[ -n '' ]] 00:13:20.081 20:12:15 -- common/autobuild_common.sh@485 -- $ '[' -n '' ']' 00:13:20.081 20:12:15 -- common/autobuild_common.sh@488 -- $ scanbuild_exclude='--exclude /home/vagrant/spdk_repo/spdk/dpdk/' 00:13:20.081 20:12:15 -- common/autobuild_common.sh@492 -- $ scanbuild_exclude+=' --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp' 00:13:20.081 20:12:15 -- common/autobuild_common.sh@494 -- $ scanbuild='scan-build -o /home/vagrant/spdk_repo/spdk/../output/scan-build-tmp --exclude /home/vagrant/spdk_repo/spdk/dpdk/ --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp --status-bugs' 00:13:20.081 20:12:15 -- common/autobuild_common.sh@495 -- $ get_config_params 00:13:20.081 20:12:15 -- common/autotest_common.sh@407 -- $ xtrace_disable 00:13:20.081 20:12:15 -- common/autotest_common.sh@10 -- $ set +x 00:13:20.081 20:12:15 -- common/autobuild_common.sh@495 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f' 00:13:20.081 20:12:15 -- common/autobuild_common.sh@497 -- $ start_monitor_resources 00:13:20.081 20:12:15 -- pm/common@17 -- $ local monitor 00:13:20.081 20:12:15 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:13:20.081 20:12:15 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:13:20.081 20:12:15 -- pm/common@21 -- $ date +%s 00:13:20.081 20:12:15 -- pm/common@25 -- $ sleep 1 00:13:20.081 20:12:15 -- pm/common@21 -- $ date +%s 00:13:20.081 20:12:15 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1727813535 00:13:20.081 20:12:15 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1727813535 00:13:20.081 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1727813535_collect-cpu-load.pm.log 00:13:20.081 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1727813535_collect-vmstat.pm.log 00:13:21.016 20:12:16 -- common/autobuild_common.sh@498 -- $ trap stop_monitor_resources EXIT 00:13:21.016 20:12:16 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:13:21.016 20:12:16 -- spdk/autobuild.sh@12 -- $ umask 022 00:13:21.016 20:12:16 -- spdk/autobuild.sh@13 -- $ cd /home/vagrant/spdk_repo/spdk 00:13:21.016 20:12:16 -- spdk/autobuild.sh@16 -- $ date -u 00:13:21.016 Tue Oct 1 08:12:16 PM UTC 2024 00:13:21.016 20:12:16 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:13:21.016 v25.01-pre-28-g1b1c3081e 00:13:21.016 20:12:16 -- spdk/autobuild.sh@19 -- $ '[' 1 -eq 1 ']' 00:13:21.016 20:12:16 -- spdk/autobuild.sh@20 -- $ run_test asan echo 'using asan' 00:13:21.016 20:12:16 -- common/autotest_common.sh@1101 -- $ '[' 3 -le 1 ']' 00:13:21.016 20:12:16 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:13:21.016 20:12:16 -- common/autotest_common.sh@10 -- $ set +x 00:13:21.016 ************************************ 00:13:21.016 START TEST asan 00:13:21.016 ************************************ 00:13:21.016 using asan 00:13:21.016 20:12:16 asan -- common/autotest_common.sh@1125 -- $ echo 'using asan' 00:13:21.016 00:13:21.016 real 0m0.000s 00:13:21.016 user 0m0.000s 00:13:21.016 sys 0m0.000s 00:13:21.016 20:12:16 asan -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:13:21.016 ************************************ 00:13:21.016 END TEST asan 00:13:21.016 20:12:16 asan -- common/autotest_common.sh@10 -- $ set +x 00:13:21.016 ************************************ 00:13:21.016 20:12:16 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:13:21.016 20:12:16 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:13:21.016 20:12:16 -- common/autotest_common.sh@1101 -- $ '[' 3 -le 1 ']' 00:13:21.016 20:12:16 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:13:21.016 20:12:16 -- common/autotest_common.sh@10 -- $ set +x 00:13:21.016 ************************************ 00:13:21.016 START TEST ubsan 00:13:21.016 ************************************ 00:13:21.016 using ubsan 00:13:21.016 20:12:16 ubsan -- common/autotest_common.sh@1125 -- $ echo 'using ubsan' 00:13:21.016 00:13:21.016 real 0m0.000s 00:13:21.016 user 0m0.000s 00:13:21.016 sys 0m0.000s 00:13:21.016 20:12:16 ubsan -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:13:21.016 ************************************ 00:13:21.016 END TEST ubsan 00:13:21.016 ************************************ 00:13:21.016 20:12:16 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:13:21.016 20:12:16 -- spdk/autobuild.sh@27 -- $ '[' -n '' ']' 00:13:21.016 20:12:16 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:13:21.016 20:12:16 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:13:21.016 20:12:16 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:13:21.016 20:12:16 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:13:21.016 20:12:16 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:13:21.016 20:12:16 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:13:21.016 20:12:16 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:13:21.016 20:12:16 -- spdk/autobuild.sh@67 -- $ /home/vagrant/spdk_repo/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-shared 00:13:21.274 Using default SPDK env in /home/vagrant/spdk_repo/spdk/lib/env_dpdk 00:13:21.274 Using default DPDK in /home/vagrant/spdk_repo/spdk/dpdk/build 00:13:21.841 Using 'verbs' RDMA provider 00:13:37.751 Configuring ISA-L (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal.log)...done. 00:13:50.024 Configuring ISA-L-crypto (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal-crypto.log)...done. 00:13:50.024 Creating mk/config.mk...done. 00:13:50.024 Creating mk/cc.flags.mk...done. 00:13:50.024 Type 'make' to build. 00:13:50.024 20:12:44 -- spdk/autobuild.sh@70 -- $ run_test make make -j10 00:13:50.024 20:12:44 -- common/autotest_common.sh@1101 -- $ '[' 3 -le 1 ']' 00:13:50.024 20:12:44 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:13:50.024 20:12:44 -- common/autotest_common.sh@10 -- $ set +x 00:13:50.024 ************************************ 00:13:50.024 START TEST make 00:13:50.024 ************************************ 00:13:50.024 20:12:44 make -- common/autotest_common.sh@1125 -- $ make -j10 00:13:50.024 make[1]: Nothing to be done for 'all'. 00:14:04.896 The Meson build system 00:14:04.896 Version: 1.5.0 00:14:04.896 Source dir: /home/vagrant/spdk_repo/spdk/dpdk 00:14:04.896 Build dir: /home/vagrant/spdk_repo/spdk/dpdk/build-tmp 00:14:04.896 Build type: native build 00:14:04.896 Program cat found: YES (/usr/bin/cat) 00:14:04.896 Project name: DPDK 00:14:04.896 Project version: 24.03.0 00:14:04.896 C compiler for the host machine: cc (gcc 13.3.1 "cc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:14:04.896 C linker for the host machine: cc ld.bfd 2.40-14 00:14:04.896 Host machine cpu family: x86_64 00:14:04.896 Host machine cpu: x86_64 00:14:04.896 Message: ## Building in Developer Mode ## 00:14:04.896 Program pkg-config found: YES (/usr/bin/pkg-config) 00:14:04.896 Program check-symbols.sh found: YES (/home/vagrant/spdk_repo/spdk/dpdk/buildtools/check-symbols.sh) 00:14:04.896 Program options-ibverbs-static.sh found: YES (/home/vagrant/spdk_repo/spdk/dpdk/buildtools/options-ibverbs-static.sh) 00:14:04.896 Program python3 found: YES (/usr/bin/python3) 00:14:04.896 Program cat found: YES (/usr/bin/cat) 00:14:04.896 Compiler for C supports arguments -march=native: YES 00:14:04.896 Checking for size of "void *" : 8 00:14:04.896 Checking for size of "void *" : 8 (cached) 00:14:04.896 Compiler for C supports link arguments -Wl,--undefined-version: YES 00:14:04.896 Library m found: YES 00:14:04.896 Library numa found: YES 00:14:04.896 Has header "numaif.h" : YES 00:14:04.896 Library fdt found: NO 00:14:04.896 Library execinfo found: NO 00:14:04.896 Has header "execinfo.h" : YES 00:14:04.896 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:14:04.896 Run-time dependency libarchive found: NO (tried pkgconfig) 00:14:04.896 Run-time dependency libbsd found: NO (tried pkgconfig) 00:14:04.896 Run-time dependency jansson found: NO (tried pkgconfig) 00:14:04.896 Run-time dependency openssl found: YES 3.1.1 00:14:04.896 Run-time dependency libpcap found: YES 1.10.4 00:14:04.896 Has header "pcap.h" with dependency libpcap: YES 00:14:04.896 Compiler for C supports arguments -Wcast-qual: YES 00:14:04.896 Compiler for C supports arguments -Wdeprecated: YES 00:14:04.896 Compiler for C supports arguments -Wformat: YES 00:14:04.896 Compiler for C supports arguments -Wformat-nonliteral: NO 00:14:04.896 Compiler for C supports arguments -Wformat-security: NO 00:14:04.896 Compiler for C supports arguments -Wmissing-declarations: YES 00:14:04.897 Compiler for C supports arguments -Wmissing-prototypes: YES 00:14:04.897 Compiler for C supports arguments -Wnested-externs: YES 00:14:04.897 Compiler for C supports arguments -Wold-style-definition: YES 00:14:04.897 Compiler for C supports arguments -Wpointer-arith: YES 00:14:04.897 Compiler for C supports arguments -Wsign-compare: YES 00:14:04.897 Compiler for C supports arguments -Wstrict-prototypes: YES 00:14:04.897 Compiler for C supports arguments -Wundef: YES 00:14:04.897 Compiler for C supports arguments -Wwrite-strings: YES 00:14:04.897 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:14:04.897 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:14:04.897 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:14:04.897 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:14:04.897 Program objdump found: YES (/usr/bin/objdump) 00:14:04.897 Compiler for C supports arguments -mavx512f: YES 00:14:04.897 Checking if "AVX512 checking" compiles: YES 00:14:04.897 Fetching value of define "__SSE4_2__" : 1 00:14:04.897 Fetching value of define "__AES__" : 1 00:14:04.897 Fetching value of define "__AVX__" : 1 00:14:04.897 Fetching value of define "__AVX2__" : 1 00:14:04.897 Fetching value of define "__AVX512BW__" : (undefined) 00:14:04.897 Fetching value of define "__AVX512CD__" : (undefined) 00:14:04.897 Fetching value of define "__AVX512DQ__" : (undefined) 00:14:04.897 Fetching value of define "__AVX512F__" : (undefined) 00:14:04.897 Fetching value of define "__AVX512VL__" : (undefined) 00:14:04.897 Fetching value of define "__PCLMUL__" : 1 00:14:04.897 Fetching value of define "__RDRND__" : 1 00:14:04.897 Fetching value of define "__RDSEED__" : 1 00:14:04.897 Fetching value of define "__VPCLMULQDQ__" : (undefined) 00:14:04.897 Fetching value of define "__znver1__" : (undefined) 00:14:04.897 Fetching value of define "__znver2__" : (undefined) 00:14:04.897 Fetching value of define "__znver3__" : (undefined) 00:14:04.897 Fetching value of define "__znver4__" : (undefined) 00:14:04.897 Library asan found: YES 00:14:04.897 Compiler for C supports arguments -Wno-format-truncation: YES 00:14:04.897 Message: lib/log: Defining dependency "log" 00:14:04.897 Message: lib/kvargs: Defining dependency "kvargs" 00:14:04.897 Message: lib/telemetry: Defining dependency "telemetry" 00:14:04.897 Library rt found: YES 00:14:04.897 Checking for function "getentropy" : NO 00:14:04.897 Message: lib/eal: Defining dependency "eal" 00:14:04.897 Message: lib/ring: Defining dependency "ring" 00:14:04.897 Message: lib/rcu: Defining dependency "rcu" 00:14:04.897 Message: lib/mempool: Defining dependency "mempool" 00:14:04.897 Message: lib/mbuf: Defining dependency "mbuf" 00:14:04.897 Fetching value of define "__PCLMUL__" : 1 (cached) 00:14:04.897 Fetching value of define "__AVX512F__" : (undefined) (cached) 00:14:04.897 Compiler for C supports arguments -mpclmul: YES 00:14:04.897 Compiler for C supports arguments -maes: YES 00:14:04.897 Compiler for C supports arguments -mavx512f: YES (cached) 00:14:04.897 Compiler for C supports arguments -mavx512bw: YES 00:14:04.897 Compiler for C supports arguments -mavx512dq: YES 00:14:04.897 Compiler for C supports arguments -mavx512vl: YES 00:14:04.897 Compiler for C supports arguments -mvpclmulqdq: YES 00:14:04.897 Compiler for C supports arguments -mavx2: YES 00:14:04.897 Compiler for C supports arguments -mavx: YES 00:14:04.897 Message: lib/net: Defining dependency "net" 00:14:04.897 Message: lib/meter: Defining dependency "meter" 00:14:04.897 Message: lib/ethdev: Defining dependency "ethdev" 00:14:04.897 Message: lib/pci: Defining dependency "pci" 00:14:04.897 Message: lib/cmdline: Defining dependency "cmdline" 00:14:04.897 Message: lib/hash: Defining dependency "hash" 00:14:04.897 Message: lib/timer: Defining dependency "timer" 00:14:04.897 Message: lib/compressdev: Defining dependency "compressdev" 00:14:04.897 Message: lib/cryptodev: Defining dependency "cryptodev" 00:14:04.897 Message: lib/dmadev: Defining dependency "dmadev" 00:14:04.897 Compiler for C supports arguments -Wno-cast-qual: YES 00:14:04.897 Message: lib/power: Defining dependency "power" 00:14:04.897 Message: lib/reorder: Defining dependency "reorder" 00:14:04.897 Message: lib/security: Defining dependency "security" 00:14:04.897 Has header "linux/userfaultfd.h" : YES 00:14:04.897 Has header "linux/vduse.h" : YES 00:14:04.897 Message: lib/vhost: Defining dependency "vhost" 00:14:04.897 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:14:04.897 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:14:04.897 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:14:04.897 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:14:04.897 Message: Disabling raw/* drivers: missing internal dependency "rawdev" 00:14:04.897 Message: Disabling regex/* drivers: missing internal dependency "regexdev" 00:14:04.897 Message: Disabling ml/* drivers: missing internal dependency "mldev" 00:14:04.897 Message: Disabling event/* drivers: missing internal dependency "eventdev" 00:14:04.897 Message: Disabling baseband/* drivers: missing internal dependency "bbdev" 00:14:04.897 Message: Disabling gpu/* drivers: missing internal dependency "gpudev" 00:14:04.897 Program doxygen found: YES (/usr/local/bin/doxygen) 00:14:04.897 Configuring doxy-api-html.conf using configuration 00:14:04.897 Configuring doxy-api-man.conf using configuration 00:14:04.897 Program mandb found: YES (/usr/bin/mandb) 00:14:04.897 Program sphinx-build found: NO 00:14:04.897 Configuring rte_build_config.h using configuration 00:14:04.897 Message: 00:14:04.897 ================= 00:14:04.897 Applications Enabled 00:14:04.897 ================= 00:14:04.897 00:14:04.897 apps: 00:14:04.897 00:14:04.897 00:14:04.897 Message: 00:14:04.897 ================= 00:14:04.897 Libraries Enabled 00:14:04.897 ================= 00:14:04.897 00:14:04.897 libs: 00:14:04.897 log, kvargs, telemetry, eal, ring, rcu, mempool, mbuf, 00:14:04.897 net, meter, ethdev, pci, cmdline, hash, timer, compressdev, 00:14:04.897 cryptodev, dmadev, power, reorder, security, vhost, 00:14:04.897 00:14:04.897 Message: 00:14:04.897 =============== 00:14:04.897 Drivers Enabled 00:14:04.897 =============== 00:14:04.897 00:14:04.897 common: 00:14:04.897 00:14:04.897 bus: 00:14:04.897 pci, vdev, 00:14:04.897 mempool: 00:14:04.897 ring, 00:14:04.897 dma: 00:14:04.897 00:14:04.897 net: 00:14:04.897 00:14:04.897 crypto: 00:14:04.897 00:14:04.897 compress: 00:14:04.897 00:14:04.897 vdpa: 00:14:04.897 00:14:04.897 00:14:04.897 Message: 00:14:04.897 ================= 00:14:04.897 Content Skipped 00:14:04.897 ================= 00:14:04.897 00:14:04.897 apps: 00:14:04.897 dumpcap: explicitly disabled via build config 00:14:04.897 graph: explicitly disabled via build config 00:14:04.897 pdump: explicitly disabled via build config 00:14:04.897 proc-info: explicitly disabled via build config 00:14:04.897 test-acl: explicitly disabled via build config 00:14:04.897 test-bbdev: explicitly disabled via build config 00:14:04.897 test-cmdline: explicitly disabled via build config 00:14:04.897 test-compress-perf: explicitly disabled via build config 00:14:04.897 test-crypto-perf: explicitly disabled via build config 00:14:04.897 test-dma-perf: explicitly disabled via build config 00:14:04.897 test-eventdev: explicitly disabled via build config 00:14:04.897 test-fib: explicitly disabled via build config 00:14:04.897 test-flow-perf: explicitly disabled via build config 00:14:04.897 test-gpudev: explicitly disabled via build config 00:14:04.897 test-mldev: explicitly disabled via build config 00:14:04.897 test-pipeline: explicitly disabled via build config 00:14:04.897 test-pmd: explicitly disabled via build config 00:14:04.897 test-regex: explicitly disabled via build config 00:14:04.897 test-sad: explicitly disabled via build config 00:14:04.897 test-security-perf: explicitly disabled via build config 00:14:04.897 00:14:04.897 libs: 00:14:04.897 argparse: explicitly disabled via build config 00:14:04.897 metrics: explicitly disabled via build config 00:14:04.897 acl: explicitly disabled via build config 00:14:04.897 bbdev: explicitly disabled via build config 00:14:04.897 bitratestats: explicitly disabled via build config 00:14:04.897 bpf: explicitly disabled via build config 00:14:04.897 cfgfile: explicitly disabled via build config 00:14:04.898 distributor: explicitly disabled via build config 00:14:04.898 efd: explicitly disabled via build config 00:14:04.898 eventdev: explicitly disabled via build config 00:14:04.898 dispatcher: explicitly disabled via build config 00:14:04.898 gpudev: explicitly disabled via build config 00:14:04.898 gro: explicitly disabled via build config 00:14:04.898 gso: explicitly disabled via build config 00:14:04.898 ip_frag: explicitly disabled via build config 00:14:04.898 jobstats: explicitly disabled via build config 00:14:04.898 latencystats: explicitly disabled via build config 00:14:04.898 lpm: explicitly disabled via build config 00:14:04.898 member: explicitly disabled via build config 00:14:04.898 pcapng: explicitly disabled via build config 00:14:04.898 rawdev: explicitly disabled via build config 00:14:04.898 regexdev: explicitly disabled via build config 00:14:04.898 mldev: explicitly disabled via build config 00:14:04.898 rib: explicitly disabled via build config 00:14:04.898 sched: explicitly disabled via build config 00:14:04.898 stack: explicitly disabled via build config 00:14:04.898 ipsec: explicitly disabled via build config 00:14:04.898 pdcp: explicitly disabled via build config 00:14:04.898 fib: explicitly disabled via build config 00:14:04.898 port: explicitly disabled via build config 00:14:04.898 pdump: explicitly disabled via build config 00:14:04.898 table: explicitly disabled via build config 00:14:04.898 pipeline: explicitly disabled via build config 00:14:04.898 graph: explicitly disabled via build config 00:14:04.898 node: explicitly disabled via build config 00:14:04.898 00:14:04.898 drivers: 00:14:04.898 common/cpt: not in enabled drivers build config 00:14:04.898 common/dpaax: not in enabled drivers build config 00:14:04.898 common/iavf: not in enabled drivers build config 00:14:04.898 common/idpf: not in enabled drivers build config 00:14:04.898 common/ionic: not in enabled drivers build config 00:14:04.898 common/mvep: not in enabled drivers build config 00:14:04.898 common/octeontx: not in enabled drivers build config 00:14:04.898 bus/auxiliary: not in enabled drivers build config 00:14:04.898 bus/cdx: not in enabled drivers build config 00:14:04.898 bus/dpaa: not in enabled drivers build config 00:14:04.898 bus/fslmc: not in enabled drivers build config 00:14:04.898 bus/ifpga: not in enabled drivers build config 00:14:04.898 bus/platform: not in enabled drivers build config 00:14:04.898 bus/uacce: not in enabled drivers build config 00:14:04.898 bus/vmbus: not in enabled drivers build config 00:14:04.898 common/cnxk: not in enabled drivers build config 00:14:04.898 common/mlx5: not in enabled drivers build config 00:14:04.898 common/nfp: not in enabled drivers build config 00:14:04.898 common/nitrox: not in enabled drivers build config 00:14:04.898 common/qat: not in enabled drivers build config 00:14:04.898 common/sfc_efx: not in enabled drivers build config 00:14:04.898 mempool/bucket: not in enabled drivers build config 00:14:04.898 mempool/cnxk: not in enabled drivers build config 00:14:04.898 mempool/dpaa: not in enabled drivers build config 00:14:04.898 mempool/dpaa2: not in enabled drivers build config 00:14:04.898 mempool/octeontx: not in enabled drivers build config 00:14:04.898 mempool/stack: not in enabled drivers build config 00:14:04.898 dma/cnxk: not in enabled drivers build config 00:14:04.898 dma/dpaa: not in enabled drivers build config 00:14:04.898 dma/dpaa2: not in enabled drivers build config 00:14:04.898 dma/hisilicon: not in enabled drivers build config 00:14:04.898 dma/idxd: not in enabled drivers build config 00:14:04.898 dma/ioat: not in enabled drivers build config 00:14:04.898 dma/skeleton: not in enabled drivers build config 00:14:04.898 net/af_packet: not in enabled drivers build config 00:14:04.898 net/af_xdp: not in enabled drivers build config 00:14:04.898 net/ark: not in enabled drivers build config 00:14:04.898 net/atlantic: not in enabled drivers build config 00:14:04.898 net/avp: not in enabled drivers build config 00:14:04.898 net/axgbe: not in enabled drivers build config 00:14:04.898 net/bnx2x: not in enabled drivers build config 00:14:04.898 net/bnxt: not in enabled drivers build config 00:14:04.898 net/bonding: not in enabled drivers build config 00:14:04.898 net/cnxk: not in enabled drivers build config 00:14:04.898 net/cpfl: not in enabled drivers build config 00:14:04.898 net/cxgbe: not in enabled drivers build config 00:14:04.898 net/dpaa: not in enabled drivers build config 00:14:04.898 net/dpaa2: not in enabled drivers build config 00:14:04.898 net/e1000: not in enabled drivers build config 00:14:04.898 net/ena: not in enabled drivers build config 00:14:04.898 net/enetc: not in enabled drivers build config 00:14:04.898 net/enetfec: not in enabled drivers build config 00:14:04.898 net/enic: not in enabled drivers build config 00:14:04.898 net/failsafe: not in enabled drivers build config 00:14:04.898 net/fm10k: not in enabled drivers build config 00:14:04.898 net/gve: not in enabled drivers build config 00:14:04.898 net/hinic: not in enabled drivers build config 00:14:04.898 net/hns3: not in enabled drivers build config 00:14:04.898 net/i40e: not in enabled drivers build config 00:14:04.898 net/iavf: not in enabled drivers build config 00:14:04.898 net/ice: not in enabled drivers build config 00:14:04.898 net/idpf: not in enabled drivers build config 00:14:04.898 net/igc: not in enabled drivers build config 00:14:04.898 net/ionic: not in enabled drivers build config 00:14:04.898 net/ipn3ke: not in enabled drivers build config 00:14:04.898 net/ixgbe: not in enabled drivers build config 00:14:04.898 net/mana: not in enabled drivers build config 00:14:04.898 net/memif: not in enabled drivers build config 00:14:04.898 net/mlx4: not in enabled drivers build config 00:14:04.898 net/mlx5: not in enabled drivers build config 00:14:04.898 net/mvneta: not in enabled drivers build config 00:14:04.898 net/mvpp2: not in enabled drivers build config 00:14:04.898 net/netvsc: not in enabled drivers build config 00:14:04.898 net/nfb: not in enabled drivers build config 00:14:04.898 net/nfp: not in enabled drivers build config 00:14:04.898 net/ngbe: not in enabled drivers build config 00:14:04.898 net/null: not in enabled drivers build config 00:14:04.898 net/octeontx: not in enabled drivers build config 00:14:04.898 net/octeon_ep: not in enabled drivers build config 00:14:04.898 net/pcap: not in enabled drivers build config 00:14:04.898 net/pfe: not in enabled drivers build config 00:14:04.898 net/qede: not in enabled drivers build config 00:14:04.898 net/ring: not in enabled drivers build config 00:14:04.898 net/sfc: not in enabled drivers build config 00:14:04.898 net/softnic: not in enabled drivers build config 00:14:04.898 net/tap: not in enabled drivers build config 00:14:04.898 net/thunderx: not in enabled drivers build config 00:14:04.898 net/txgbe: not in enabled drivers build config 00:14:04.898 net/vdev_netvsc: not in enabled drivers build config 00:14:04.898 net/vhost: not in enabled drivers build config 00:14:04.898 net/virtio: not in enabled drivers build config 00:14:04.898 net/vmxnet3: not in enabled drivers build config 00:14:04.898 raw/*: missing internal dependency, "rawdev" 00:14:04.898 crypto/armv8: not in enabled drivers build config 00:14:04.898 crypto/bcmfs: not in enabled drivers build config 00:14:04.898 crypto/caam_jr: not in enabled drivers build config 00:14:04.898 crypto/ccp: not in enabled drivers build config 00:14:04.898 crypto/cnxk: not in enabled drivers build config 00:14:04.898 crypto/dpaa_sec: not in enabled drivers build config 00:14:04.898 crypto/dpaa2_sec: not in enabled drivers build config 00:14:04.898 crypto/ipsec_mb: not in enabled drivers build config 00:14:04.898 crypto/mlx5: not in enabled drivers build config 00:14:04.898 crypto/mvsam: not in enabled drivers build config 00:14:04.898 crypto/nitrox: not in enabled drivers build config 00:14:04.898 crypto/null: not in enabled drivers build config 00:14:04.898 crypto/octeontx: not in enabled drivers build config 00:14:04.898 crypto/openssl: not in enabled drivers build config 00:14:04.898 crypto/scheduler: not in enabled drivers build config 00:14:04.898 crypto/uadk: not in enabled drivers build config 00:14:04.898 crypto/virtio: not in enabled drivers build config 00:14:04.898 compress/isal: not in enabled drivers build config 00:14:04.898 compress/mlx5: not in enabled drivers build config 00:14:04.898 compress/nitrox: not in enabled drivers build config 00:14:04.898 compress/octeontx: not in enabled drivers build config 00:14:04.898 compress/zlib: not in enabled drivers build config 00:14:04.898 regex/*: missing internal dependency, "regexdev" 00:14:04.898 ml/*: missing internal dependency, "mldev" 00:14:04.898 vdpa/ifc: not in enabled drivers build config 00:14:04.898 vdpa/mlx5: not in enabled drivers build config 00:14:04.898 vdpa/nfp: not in enabled drivers build config 00:14:04.898 vdpa/sfc: not in enabled drivers build config 00:14:04.898 event/*: missing internal dependency, "eventdev" 00:14:04.898 baseband/*: missing internal dependency, "bbdev" 00:14:04.898 gpu/*: missing internal dependency, "gpudev" 00:14:04.898 00:14:04.898 00:14:04.898 Build targets in project: 85 00:14:04.898 00:14:04.898 DPDK 24.03.0 00:14:04.898 00:14:04.898 User defined options 00:14:04.898 buildtype : debug 00:14:04.898 default_library : shared 00:14:04.898 libdir : lib 00:14:04.899 prefix : /home/vagrant/spdk_repo/spdk/dpdk/build 00:14:04.899 b_sanitize : address 00:14:04.899 c_args : -Wno-stringop-overflow -fcommon -Wno-stringop-overread -Wno-array-bounds -fPIC -Werror 00:14:04.899 c_link_args : 00:14:04.899 cpu_instruction_set: native 00:14:04.899 disable_apps : dumpcap,graph,pdump,proc-info,test-acl,test-bbdev,test-cmdline,test-compress-perf,test-crypto-perf,test-dma-perf,test-eventdev,test-fib,test-flow-perf,test-gpudev,test-mldev,test-pipeline,test-pmd,test-regex,test-sad,test-security-perf,test 00:14:04.899 disable_libs : acl,argparse,bbdev,bitratestats,bpf,cfgfile,dispatcher,distributor,efd,eventdev,fib,gpudev,graph,gro,gso,ip_frag,ipsec,jobstats,latencystats,lpm,member,metrics,mldev,node,pcapng,pdcp,pdump,pipeline,port,rawdev,regexdev,rib,sched,stack,table 00:14:04.899 enable_docs : false 00:14:04.899 enable_drivers : bus,bus/pci,bus/vdev,mempool/ring 00:14:04.899 enable_kmods : false 00:14:04.899 max_lcores : 128 00:14:04.899 tests : false 00:14:04.899 00:14:04.899 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:14:04.899 ninja: Entering directory `/home/vagrant/spdk_repo/spdk/dpdk/build-tmp' 00:14:04.899 [1/268] Compiling C object lib/librte_log.a.p/log_log_linux.c.o 00:14:04.899 [2/268] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:14:04.899 [3/268] Linking static target lib/librte_kvargs.a 00:14:04.899 [4/268] Compiling C object lib/librte_log.a.p/log_log.c.o 00:14:04.899 [5/268] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:14:04.899 [6/268] Linking static target lib/librte_log.a 00:14:04.899 [7/268] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:14:04.899 [8/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:14:04.899 [9/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:14:04.899 [10/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:14:04.899 [11/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:14:04.899 [12/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:14:04.899 [13/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:14:04.899 [14/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:14:04.899 [15/268] Generating lib/log.sym_chk with a custom command (wrapped by meson to capture output) 00:14:04.899 [16/268] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:14:04.899 [17/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:14:04.899 [18/268] Linking target lib/librte_log.so.24.1 00:14:04.899 [19/268] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:14:04.899 [20/268] Linking static target lib/librte_telemetry.a 00:14:04.899 [21/268] Generating symbol file lib/librte_log.so.24.1.p/librte_log.so.24.1.symbols 00:14:05.157 [22/268] Linking target lib/librte_kvargs.so.24.1 00:14:05.415 [23/268] Generating symbol file lib/librte_kvargs.so.24.1.p/librte_kvargs.so.24.1.symbols 00:14:05.415 [24/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:14:05.415 [25/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:14:05.415 [26/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:14:05.415 [27/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:14:05.674 [28/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:14:05.674 [29/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:14:05.674 [30/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:14:05.674 [31/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:14:05.674 [32/268] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:14:05.674 [33/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:14:05.935 [34/268] Linking target lib/librte_telemetry.so.24.1 00:14:06.193 [35/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:14:06.194 [36/268] Generating symbol file lib/librte_telemetry.so.24.1.p/librte_telemetry.so.24.1.symbols 00:14:06.194 [37/268] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:14:06.452 [38/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:14:06.452 [39/268] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:14:06.452 [40/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:14:06.452 [41/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:14:06.452 [42/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:14:06.452 [43/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:14:06.711 [44/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:14:06.711 [45/268] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:14:06.711 [46/268] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:14:06.969 [47/268] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:14:06.969 [48/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:14:07.227 [49/268] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:14:07.227 [50/268] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:14:07.227 [51/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:14:07.485 [52/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:14:07.485 [53/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:14:07.485 [54/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:14:07.485 [55/268] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:14:07.742 [56/268] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:14:07.742 [57/268] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:14:07.999 [58/268] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:14:07.999 [59/268] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:14:07.999 [60/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:14:07.999 [61/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:14:08.257 [62/268] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:14:08.257 [63/268] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:14:08.515 [64/268] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:14:08.515 [65/268] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:14:08.515 [66/268] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:14:08.773 [67/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:14:08.773 [68/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:14:08.773 [69/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:14:08.773 [70/268] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:14:09.031 [71/268] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:14:09.031 [72/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:14:09.031 [73/268] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:14:09.031 [74/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:14:09.031 [75/268] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:14:09.289 [76/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:14:09.289 [77/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:14:09.289 [78/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:14:09.547 [79/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:14:09.547 [80/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:14:09.547 [81/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:14:09.805 [82/268] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:14:09.805 [83/268] Linking static target lib/librte_ring.a 00:14:09.805 [84/268] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:14:09.805 [85/268] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:14:09.805 [86/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:14:10.063 [87/268] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:14:10.063 [88/268] Linking static target lib/librte_eal.a 00:14:10.321 [89/268] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:14:10.321 [90/268] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:14:10.321 [91/268] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:14:10.321 [92/268] Linking static target lib/librte_rcu.a 00:14:10.321 [93/268] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:14:10.579 [94/268] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:14:10.579 [95/268] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:14:10.579 [96/268] Compiling C object lib/net/libnet_crc_avx512_lib.a.p/net_crc_avx512.c.o 00:14:10.579 [97/268] Linking static target lib/net/libnet_crc_avx512_lib.a 00:14:10.837 [98/268] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:14:10.837 [99/268] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:14:10.837 [100/268] Linking static target lib/librte_mempool.a 00:14:10.837 [101/268] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:14:11.096 [102/268] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:14:11.096 [103/268] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:14:11.096 [104/268] Linking static target lib/librte_mbuf.a 00:14:11.354 [105/268] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:14:11.354 [106/268] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:14:11.612 [107/268] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:14:11.612 [108/268] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:14:11.612 [109/268] Linking static target lib/librte_meter.a 00:14:11.612 [110/268] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:14:11.612 [111/268] Linking static target lib/librte_net.a 00:14:11.612 [112/268] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:14:11.871 [113/268] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:14:11.871 [114/268] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:14:12.130 [115/268] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:14:12.130 [116/268] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:14:12.130 [117/268] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:14:12.462 [118/268] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:14:12.462 [119/268] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:14:12.722 [120/268] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:14:12.981 [121/268] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:14:12.981 [122/268] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:14:13.240 [123/268] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_telemetry.c.o 00:14:13.499 [124/268] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:14:13.499 [125/268] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:14:13.499 [126/268] Linking static target lib/librte_pci.a 00:14:13.499 [127/268] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:14:13.758 [128/268] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:14:13.758 [129/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:14:13.758 [130/268] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:14:13.758 [131/268] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:14:13.758 [132/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:14:13.758 [133/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:14:14.015 [134/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:14:14.015 [135/268] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:14:14.015 [136/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:14:14.015 [137/268] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:14:14.015 [138/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:14:14.015 [139/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:14:14.015 [140/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:14:14.015 [141/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:14:14.273 [142/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:14:14.273 [143/268] Compiling C object lib/librte_hash.a.p/hash_rte_hash_crc.c.o 00:14:14.273 [144/268] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_linux_ethtool.c.o 00:14:14.273 [145/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:14:14.273 [146/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:14:14.531 [147/268] Linking static target lib/librte_cmdline.a 00:14:15.098 [148/268] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:14:15.098 [149/268] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:14:15.098 [150/268] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:14:15.098 [151/268] Linking static target lib/librte_timer.a 00:14:15.098 [152/268] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:14:15.098 [153/268] Compiling C object lib/librte_hash.a.p/hash_rte_thash_gfni.c.o 00:14:15.098 [154/268] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:14:15.356 [155/268] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:14:15.614 [156/268] Linking static target lib/librte_ethdev.a 00:14:15.614 [157/268] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:14:15.614 [158/268] Linking static target lib/librte_hash.a 00:14:15.614 [159/268] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:14:15.873 [160/268] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:14:15.873 [161/268] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:14:15.873 [162/268] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:14:15.873 [163/268] Linking static target lib/librte_compressdev.a 00:14:15.873 [164/268] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:14:16.201 [165/268] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:14:16.201 [166/268] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:14:16.462 [167/268] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:14:16.462 [168/268] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev_trace_points.c.o 00:14:16.462 [169/268] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:14:16.462 [170/268] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:14:16.462 [171/268] Linking static target lib/librte_dmadev.a 00:14:16.720 [172/268] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:14:16.720 [173/268] Compiling C object lib/librte_power.a.p/power_power_amd_pstate_cpufreq.c.o 00:14:16.979 [174/268] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:14:16.979 [175/268] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:14:16.979 [176/268] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:14:17.237 [177/268] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:14:17.237 [178/268] Compiling C object lib/librte_power.a.p/power_power_intel_uncore.c.o 00:14:17.237 [179/268] Linking static target lib/librte_cryptodev.a 00:14:17.497 [180/268] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:14:17.497 [181/268] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:14:17.497 [182/268] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:14:17.497 [183/268] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:14:17.497 [184/268] Compiling C object lib/librte_power.a.p/power_rte_power_uncore.c.o 00:14:18.063 [185/268] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:14:18.063 [186/268] Linking static target lib/librte_power.a 00:14:18.322 [187/268] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:14:18.322 [188/268] Linking static target lib/librte_reorder.a 00:14:18.322 [189/268] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:14:18.322 [190/268] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:14:18.322 [191/268] Linking static target lib/librte_security.a 00:14:18.322 [192/268] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:14:18.581 [193/268] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:14:18.839 [194/268] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:14:18.839 [195/268] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:14:19.405 [196/268] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:14:19.405 [197/268] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:14:19.405 [198/268] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:14:19.664 [199/268] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net_ctrl.c.o 00:14:19.664 [200/268] Compiling C object lib/librte_vhost.a.p/vhost_vduse.c.o 00:14:19.922 [201/268] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:14:19.922 [202/268] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:14:20.184 [203/268] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:14:20.184 [204/268] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:14:20.445 [205/268] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:14:20.445 [206/268] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:14:20.445 [207/268] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:14:20.704 [208/268] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:14:20.704 [209/268] Linking static target drivers/libtmp_rte_bus_vdev.a 00:14:20.964 [210/268] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:14:20.964 [211/268] Linking static target drivers/libtmp_rte_bus_pci.a 00:14:20.964 [212/268] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:14:20.964 [213/268] Compiling C object drivers/librte_bus_vdev.so.24.1.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:14:20.964 [214/268] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:14:20.964 [215/268] Linking static target drivers/librte_bus_vdev.a 00:14:21.224 [216/268] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:14:21.224 [217/268] Compiling C object drivers/librte_bus_pci.so.24.1.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:14:21.224 [218/268] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:14:21.224 [219/268] Linking static target drivers/librte_bus_pci.a 00:14:21.224 [220/268] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:14:21.224 [221/268] Linking static target drivers/libtmp_rte_mempool_ring.a 00:14:21.483 [222/268] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:14:21.483 [223/268] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:14:21.483 [224/268] Compiling C object drivers/librte_mempool_ring.so.24.1.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:14:21.483 [225/268] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:14:21.483 [226/268] Linking static target drivers/librte_mempool_ring.a 00:14:21.742 [227/268] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:14:22.679 [228/268] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:14:22.679 [229/268] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:14:22.679 [230/268] Linking target lib/librte_eal.so.24.1 00:14:22.938 [231/268] Generating symbol file lib/librte_eal.so.24.1.p/librte_eal.so.24.1.symbols 00:14:22.938 [232/268] Linking target lib/librte_meter.so.24.1 00:14:22.938 [233/268] Linking target lib/librte_pci.so.24.1 00:14:22.938 [234/268] Linking target lib/librte_timer.so.24.1 00:14:22.938 [235/268] Linking target drivers/librte_bus_vdev.so.24.1 00:14:22.938 [236/268] Linking target lib/librte_ring.so.24.1 00:14:22.938 [237/268] Linking target lib/librte_dmadev.so.24.1 00:14:23.197 [238/268] Generating symbol file lib/librte_pci.so.24.1.p/librte_pci.so.24.1.symbols 00:14:23.197 [239/268] Generating symbol file lib/librte_timer.so.24.1.p/librte_timer.so.24.1.symbols 00:14:23.197 [240/268] Generating symbol file lib/librte_meter.so.24.1.p/librte_meter.so.24.1.symbols 00:14:23.197 [241/268] Linking target drivers/librte_bus_pci.so.24.1 00:14:23.197 [242/268] Generating symbol file lib/librte_dmadev.so.24.1.p/librte_dmadev.so.24.1.symbols 00:14:23.197 [243/268] Generating symbol file lib/librte_ring.so.24.1.p/librte_ring.so.24.1.symbols 00:14:23.197 [244/268] Linking target lib/librte_rcu.so.24.1 00:14:23.197 [245/268] Linking target lib/librte_mempool.so.24.1 00:14:23.456 [246/268] Generating symbol file lib/librte_mempool.so.24.1.p/librte_mempool.so.24.1.symbols 00:14:23.456 [247/268] Generating symbol file lib/librte_rcu.so.24.1.p/librte_rcu.so.24.1.symbols 00:14:23.456 [248/268] Linking target drivers/librte_mempool_ring.so.24.1 00:14:23.456 [249/268] Linking target lib/librte_mbuf.so.24.1 00:14:23.715 [250/268] Generating symbol file lib/librte_mbuf.so.24.1.p/librte_mbuf.so.24.1.symbols 00:14:23.715 [251/268] Linking target lib/librte_compressdev.so.24.1 00:14:23.715 [252/268] Linking target lib/librte_reorder.so.24.1 00:14:23.715 [253/268] Linking target lib/librte_net.so.24.1 00:14:23.715 [254/268] Linking target lib/librte_cryptodev.so.24.1 00:14:23.715 [255/268] Generating symbol file lib/librte_net.so.24.1.p/librte_net.so.24.1.symbols 00:14:23.715 [256/268] Generating symbol file lib/librte_cryptodev.so.24.1.p/librte_cryptodev.so.24.1.symbols 00:14:23.974 [257/268] Linking target lib/librte_hash.so.24.1 00:14:23.974 [258/268] Linking target lib/librte_cmdline.so.24.1 00:14:23.974 [259/268] Linking target lib/librte_security.so.24.1 00:14:23.974 [260/268] Generating symbol file lib/librte_hash.so.24.1.p/librte_hash.so.24.1.symbols 00:14:23.974 [261/268] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:14:24.231 [262/268] Linking target lib/librte_ethdev.so.24.1 00:14:24.490 [263/268] Generating symbol file lib/librte_ethdev.so.24.1.p/librte_ethdev.so.24.1.symbols 00:14:24.490 [264/268] Linking target lib/librte_power.so.24.1 00:14:27.021 [265/268] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:14:27.021 [266/268] Linking static target lib/librte_vhost.a 00:14:28.937 [267/268] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:14:28.937 [268/268] Linking target lib/librte_vhost.so.24.1 00:14:28.937 INFO: autodetecting backend as ninja 00:14:28.937 INFO: calculating backend command to run: /usr/local/bin/ninja -C /home/vagrant/spdk_repo/spdk/dpdk/build-tmp -j 10 00:14:50.898 CC lib/log/log.o 00:14:50.898 CC lib/ut/ut.o 00:14:50.899 CC lib/log/log_deprecated.o 00:14:50.899 CC lib/log/log_flags.o 00:14:50.899 CC lib/ut_mock/mock.o 00:14:50.899 LIB libspdk_log.a 00:14:50.899 LIB libspdk_ut.a 00:14:50.899 LIB libspdk_ut_mock.a 00:14:50.899 SO libspdk_log.so.7.0 00:14:50.899 SO libspdk_ut.so.2.0 00:14:50.899 SO libspdk_ut_mock.so.6.0 00:14:51.157 SYMLINK libspdk_ut.so 00:14:51.157 SYMLINK libspdk_ut_mock.so 00:14:51.157 SYMLINK libspdk_log.so 00:14:51.157 CXX lib/trace_parser/trace.o 00:14:51.157 CC lib/dma/dma.o 00:14:51.157 CC lib/util/bit_array.o 00:14:51.157 CC lib/util/base64.o 00:14:51.157 CC lib/util/cpuset.o 00:14:51.157 CC lib/util/crc32.o 00:14:51.157 CC lib/util/crc16.o 00:14:51.157 CC lib/util/crc32c.o 00:14:51.157 CC lib/ioat/ioat.o 00:14:51.414 CC lib/vfio_user/host/vfio_user_pci.o 00:14:51.414 CC lib/util/crc32_ieee.o 00:14:51.414 CC lib/util/crc64.o 00:14:51.414 CC lib/util/dif.o 00:14:51.414 CC lib/util/fd.o 00:14:51.710 LIB libspdk_dma.a 00:14:51.710 CC lib/util/fd_group.o 00:14:51.710 SO libspdk_dma.so.5.0 00:14:51.710 CC lib/util/file.o 00:14:51.710 CC lib/util/hexlify.o 00:14:51.710 CC lib/util/iov.o 00:14:51.710 CC lib/util/math.o 00:14:51.710 SYMLINK libspdk_dma.so 00:14:51.710 CC lib/vfio_user/host/vfio_user.o 00:14:51.710 LIB libspdk_ioat.a 00:14:51.710 CC lib/util/net.o 00:14:51.710 SO libspdk_ioat.so.7.0 00:14:51.710 CC lib/util/pipe.o 00:14:51.710 CC lib/util/strerror_tls.o 00:14:51.710 CC lib/util/string.o 00:14:51.710 SYMLINK libspdk_ioat.so 00:14:51.710 CC lib/util/uuid.o 00:14:51.983 CC lib/util/xor.o 00:14:51.983 CC lib/util/zipf.o 00:14:51.983 LIB libspdk_vfio_user.a 00:14:51.983 SO libspdk_vfio_user.so.5.0 00:14:51.983 CC lib/util/md5.o 00:14:51.983 SYMLINK libspdk_vfio_user.so 00:14:52.241 LIB libspdk_util.a 00:14:52.241 SO libspdk_util.so.10.0 00:14:52.500 SYMLINK libspdk_util.so 00:14:52.500 LIB libspdk_trace_parser.a 00:14:52.500 SO libspdk_trace_parser.so.6.0 00:14:52.759 SYMLINK libspdk_trace_parser.so 00:14:52.759 CC lib/conf/conf.o 00:14:52.759 CC lib/idxd/idxd_user.o 00:14:52.759 CC lib/idxd/idxd.o 00:14:52.759 CC lib/idxd/idxd_kernel.o 00:14:52.759 CC lib/vmd/vmd.o 00:14:52.759 CC lib/rdma_provider/common.o 00:14:52.759 CC lib/rdma_provider/rdma_provider_verbs.o 00:14:52.759 CC lib/json/json_parse.o 00:14:52.759 CC lib/env_dpdk/env.o 00:14:52.759 CC lib/rdma_utils/rdma_utils.o 00:14:52.759 CC lib/env_dpdk/memory.o 00:14:53.018 CC lib/env_dpdk/pci.o 00:14:53.018 LIB libspdk_rdma_provider.a 00:14:53.018 CC lib/json/json_util.o 00:14:53.018 SO libspdk_rdma_provider.so.6.0 00:14:53.018 LIB libspdk_conf.a 00:14:53.018 CC lib/env_dpdk/init.o 00:14:53.018 SO libspdk_conf.so.6.0 00:14:53.018 LIB libspdk_rdma_utils.a 00:14:53.018 SYMLINK libspdk_rdma_provider.so 00:14:53.018 CC lib/json/json_write.o 00:14:53.018 SYMLINK libspdk_conf.so 00:14:53.018 CC lib/env_dpdk/threads.o 00:14:53.018 SO libspdk_rdma_utils.so.1.0 00:14:53.276 SYMLINK libspdk_rdma_utils.so 00:14:53.276 CC lib/env_dpdk/pci_ioat.o 00:14:53.276 CC lib/env_dpdk/pci_virtio.o 00:14:53.276 CC lib/env_dpdk/pci_vmd.o 00:14:53.276 CC lib/env_dpdk/pci_idxd.o 00:14:53.536 CC lib/env_dpdk/pci_event.o 00:14:53.536 CC lib/env_dpdk/sigbus_handler.o 00:14:53.536 CC lib/env_dpdk/pci_dpdk.o 00:14:53.536 CC lib/env_dpdk/pci_dpdk_2207.o 00:14:53.536 LIB libspdk_json.a 00:14:53.536 CC lib/env_dpdk/pci_dpdk_2211.o 00:14:53.536 SO libspdk_json.so.6.0 00:14:53.536 SYMLINK libspdk_json.so 00:14:53.536 CC lib/vmd/led.o 00:14:53.536 LIB libspdk_idxd.a 00:14:53.795 SO libspdk_idxd.so.12.1 00:14:53.795 SYMLINK libspdk_idxd.so 00:14:53.795 LIB libspdk_vmd.a 00:14:53.795 CC lib/jsonrpc/jsonrpc_server.o 00:14:53.795 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:14:53.795 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:14:53.795 CC lib/jsonrpc/jsonrpc_client.o 00:14:53.795 SO libspdk_vmd.so.6.0 00:14:54.053 SYMLINK libspdk_vmd.so 00:14:54.053 LIB libspdk_jsonrpc.a 00:14:54.311 SO libspdk_jsonrpc.so.6.0 00:14:54.311 SYMLINK libspdk_jsonrpc.so 00:14:54.570 CC lib/rpc/rpc.o 00:14:54.829 LIB libspdk_env_dpdk.a 00:14:54.829 SO libspdk_env_dpdk.so.15.0 00:14:54.829 LIB libspdk_rpc.a 00:14:54.829 SO libspdk_rpc.so.6.0 00:14:55.089 SYMLINK libspdk_rpc.so 00:14:55.089 SYMLINK libspdk_env_dpdk.so 00:14:55.347 CC lib/keyring/keyring_rpc.o 00:14:55.347 CC lib/keyring/keyring.o 00:14:55.347 CC lib/notify/notify_rpc.o 00:14:55.347 CC lib/notify/notify.o 00:14:55.347 CC lib/trace/trace.o 00:14:55.347 CC lib/trace/trace_flags.o 00:14:55.347 CC lib/trace/trace_rpc.o 00:14:55.605 LIB libspdk_notify.a 00:14:55.605 LIB libspdk_keyring.a 00:14:55.605 SO libspdk_notify.so.6.0 00:14:55.605 SO libspdk_keyring.so.2.0 00:14:55.605 LIB libspdk_trace.a 00:14:55.605 SYMLINK libspdk_keyring.so 00:14:55.605 SYMLINK libspdk_notify.so 00:14:55.605 SO libspdk_trace.so.11.0 00:14:55.863 SYMLINK libspdk_trace.so 00:14:56.122 CC lib/sock/sock.o 00:14:56.122 CC lib/sock/sock_rpc.o 00:14:56.122 CC lib/thread/iobuf.o 00:14:56.122 CC lib/thread/thread.o 00:14:56.688 LIB libspdk_sock.a 00:14:56.688 SO libspdk_sock.so.10.0 00:14:56.688 SYMLINK libspdk_sock.so 00:14:56.946 CC lib/nvme/nvme_ctrlr_cmd.o 00:14:56.946 CC lib/nvme/nvme_ctrlr.o 00:14:56.946 CC lib/nvme/nvme_fabric.o 00:14:56.946 CC lib/nvme/nvme_ns_cmd.o 00:14:56.946 CC lib/nvme/nvme_ns.o 00:14:56.946 CC lib/nvme/nvme_pcie.o 00:14:56.946 CC lib/nvme/nvme_pcie_common.o 00:14:56.946 CC lib/nvme/nvme_qpair.o 00:14:56.946 CC lib/nvme/nvme.o 00:14:57.881 CC lib/nvme/nvme_quirks.o 00:14:57.881 CC lib/nvme/nvme_transport.o 00:14:58.139 CC lib/nvme/nvme_discovery.o 00:14:58.139 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:14:58.139 LIB libspdk_thread.a 00:14:58.139 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:14:58.139 SO libspdk_thread.so.10.1 00:14:58.139 SYMLINK libspdk_thread.so 00:14:58.139 CC lib/nvme/nvme_tcp.o 00:14:58.397 CC lib/nvme/nvme_opal.o 00:14:58.397 CC lib/nvme/nvme_io_msg.o 00:14:58.397 CC lib/nvme/nvme_poll_group.o 00:14:58.655 CC lib/nvme/nvme_zns.o 00:14:58.655 CC lib/nvme/nvme_stubs.o 00:14:58.655 CC lib/nvme/nvme_auth.o 00:14:58.913 CC lib/nvme/nvme_cuse.o 00:14:58.913 CC lib/nvme/nvme_rdma.o 00:14:59.171 CC lib/accel/accel.o 00:14:59.171 CC lib/blob/blobstore.o 00:14:59.430 CC lib/init/json_config.o 00:14:59.430 CC lib/init/subsystem.o 00:14:59.430 CC lib/init/subsystem_rpc.o 00:14:59.688 CC lib/init/rpc.o 00:14:59.688 CC lib/blob/request.o 00:14:59.688 CC lib/blob/zeroes.o 00:14:59.688 LIB libspdk_init.a 00:14:59.947 SO libspdk_init.so.6.0 00:14:59.947 CC lib/blob/blob_bs_dev.o 00:14:59.947 SYMLINK libspdk_init.so 00:14:59.947 CC lib/accel/accel_rpc.o 00:14:59.947 CC lib/accel/accel_sw.o 00:14:59.947 CC lib/virtio/virtio.o 00:15:00.205 CC lib/fsdev/fsdev.o 00:15:00.205 CC lib/fsdev/fsdev_io.o 00:15:00.205 CC lib/fsdev/fsdev_rpc.o 00:15:00.205 CC lib/virtio/virtio_vhost_user.o 00:15:00.464 CC lib/virtio/virtio_vfio_user.o 00:15:00.464 CC lib/virtio/virtio_pci.o 00:15:00.464 CC lib/event/app.o 00:15:00.464 CC lib/event/reactor.o 00:15:00.723 CC lib/event/log_rpc.o 00:15:00.723 CC lib/event/app_rpc.o 00:15:00.723 LIB libspdk_accel.a 00:15:00.723 SO libspdk_accel.so.16.0 00:15:00.723 LIB libspdk_nvme.a 00:15:00.723 LIB libspdk_virtio.a 00:15:00.723 CC lib/event/scheduler_static.o 00:15:00.982 SYMLINK libspdk_accel.so 00:15:00.982 SO libspdk_virtio.so.7.0 00:15:00.982 SYMLINK libspdk_virtio.so 00:15:00.982 LIB libspdk_fsdev.a 00:15:00.982 SO libspdk_nvme.so.14.0 00:15:00.982 SO libspdk_fsdev.so.1.0 00:15:01.240 CC lib/bdev/bdev.o 00:15:01.240 CC lib/bdev/bdev_rpc.o 00:15:01.240 CC lib/bdev/bdev_zone.o 00:15:01.240 CC lib/bdev/scsi_nvme.o 00:15:01.240 CC lib/bdev/part.o 00:15:01.240 SYMLINK libspdk_fsdev.so 00:15:01.240 LIB libspdk_event.a 00:15:01.240 SO libspdk_event.so.14.0 00:15:01.499 CC lib/fuse_dispatcher/fuse_dispatcher.o 00:15:01.499 SYMLINK libspdk_nvme.so 00:15:01.499 SYMLINK libspdk_event.so 00:15:02.435 LIB libspdk_fuse_dispatcher.a 00:15:02.435 SO libspdk_fuse_dispatcher.so.1.0 00:15:02.435 SYMLINK libspdk_fuse_dispatcher.so 00:15:04.338 LIB libspdk_blob.a 00:15:04.338 SO libspdk_blob.so.11.0 00:15:04.338 SYMLINK libspdk_blob.so 00:15:04.597 CC lib/lvol/lvol.o 00:15:04.597 CC lib/blobfs/blobfs.o 00:15:04.597 CC lib/blobfs/tree.o 00:15:05.163 LIB libspdk_bdev.a 00:15:05.163 SO libspdk_bdev.so.17.0 00:15:05.421 SYMLINK libspdk_bdev.so 00:15:05.681 CC lib/ublk/ublk.o 00:15:05.681 CC lib/ublk/ublk_rpc.o 00:15:05.681 CC lib/ftl/ftl_core.o 00:15:05.681 CC lib/ftl/ftl_init.o 00:15:05.681 CC lib/ftl/ftl_layout.o 00:15:05.681 CC lib/nbd/nbd.o 00:15:05.681 CC lib/scsi/dev.o 00:15:05.681 CC lib/nvmf/ctrlr.o 00:15:05.681 LIB libspdk_blobfs.a 00:15:05.681 SO libspdk_blobfs.so.10.0 00:15:05.939 LIB libspdk_lvol.a 00:15:05.939 CC lib/nvmf/ctrlr_discovery.o 00:15:05.939 SO libspdk_lvol.so.10.0 00:15:05.939 SYMLINK libspdk_blobfs.so 00:15:05.939 CC lib/nvmf/ctrlr_bdev.o 00:15:05.939 CC lib/scsi/lun.o 00:15:05.939 SYMLINK libspdk_lvol.so 00:15:05.939 CC lib/scsi/port.o 00:15:05.939 CC lib/scsi/scsi.o 00:15:06.198 CC lib/ftl/ftl_debug.o 00:15:06.198 CC lib/ftl/ftl_io.o 00:15:06.198 CC lib/scsi/scsi_bdev.o 00:15:06.198 CC lib/ftl/ftl_sb.o 00:15:06.198 CC lib/scsi/scsi_pr.o 00:15:06.198 CC lib/nbd/nbd_rpc.o 00:15:06.457 CC lib/scsi/scsi_rpc.o 00:15:06.457 CC lib/nvmf/subsystem.o 00:15:06.457 CC lib/scsi/task.o 00:15:06.457 CC lib/ftl/ftl_l2p.o 00:15:06.457 LIB libspdk_nbd.a 00:15:06.457 CC lib/ftl/ftl_l2p_flat.o 00:15:06.457 SO libspdk_nbd.so.7.0 00:15:06.715 SYMLINK libspdk_nbd.so 00:15:06.715 CC lib/ftl/ftl_nv_cache.o 00:15:06.715 CC lib/ftl/ftl_band.o 00:15:06.715 LIB libspdk_ublk.a 00:15:06.715 CC lib/ftl/ftl_band_ops.o 00:15:06.715 CC lib/nvmf/nvmf.o 00:15:06.715 SO libspdk_ublk.so.3.0 00:15:06.715 CC lib/nvmf/nvmf_rpc.o 00:15:06.715 SYMLINK libspdk_ublk.so 00:15:06.715 CC lib/ftl/ftl_writer.o 00:15:06.715 CC lib/ftl/ftl_rq.o 00:15:06.972 LIB libspdk_scsi.a 00:15:06.972 SO libspdk_scsi.so.9.0 00:15:06.972 CC lib/ftl/ftl_reloc.o 00:15:06.972 SYMLINK libspdk_scsi.so 00:15:06.972 CC lib/ftl/ftl_l2p_cache.o 00:15:07.231 CC lib/nvmf/transport.o 00:15:07.231 CC lib/nvmf/tcp.o 00:15:07.231 CC lib/ftl/ftl_p2l.o 00:15:07.488 CC lib/ftl/ftl_p2l_log.o 00:15:07.804 CC lib/iscsi/conn.o 00:15:07.804 CC lib/iscsi/init_grp.o 00:15:07.804 CC lib/ftl/mngt/ftl_mngt.o 00:15:07.804 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:15:07.804 CC lib/nvmf/stubs.o 00:15:08.074 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:15:08.074 CC lib/ftl/mngt/ftl_mngt_startup.o 00:15:08.074 CC lib/ftl/mngt/ftl_mngt_md.o 00:15:08.332 CC lib/iscsi/iscsi.o 00:15:08.332 CC lib/iscsi/param.o 00:15:08.332 CC lib/iscsi/portal_grp.o 00:15:08.332 CC lib/iscsi/tgt_node.o 00:15:08.332 CC lib/iscsi/iscsi_subsystem.o 00:15:08.591 CC lib/iscsi/iscsi_rpc.o 00:15:08.591 CC lib/vhost/vhost.o 00:15:08.591 CC lib/ftl/mngt/ftl_mngt_misc.o 00:15:08.591 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:15:08.591 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:15:08.591 CC lib/ftl/mngt/ftl_mngt_band.o 00:15:08.848 CC lib/vhost/vhost_rpc.o 00:15:08.848 CC lib/vhost/vhost_scsi.o 00:15:08.848 CC lib/vhost/vhost_blk.o 00:15:08.848 CC lib/vhost/rte_vhost_user.o 00:15:08.848 CC lib/iscsi/task.o 00:15:09.106 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:15:09.106 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:15:09.106 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:15:09.106 CC lib/nvmf/mdns_server.o 00:15:09.366 CC lib/nvmf/rdma.o 00:15:09.366 CC lib/nvmf/auth.o 00:15:09.366 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:15:09.624 CC lib/ftl/utils/ftl_conf.o 00:15:09.624 CC lib/ftl/utils/ftl_md.o 00:15:09.624 CC lib/ftl/utils/ftl_mempool.o 00:15:09.883 CC lib/ftl/utils/ftl_bitmap.o 00:15:09.883 CC lib/ftl/utils/ftl_property.o 00:15:09.883 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:15:09.883 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:15:10.141 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:15:10.141 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:15:10.141 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:15:10.141 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:15:10.141 LIB libspdk_vhost.a 00:15:10.141 LIB libspdk_iscsi.a 00:15:10.141 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:15:10.141 CC lib/ftl/upgrade/ftl_sb_v3.o 00:15:10.399 CC lib/ftl/upgrade/ftl_sb_v5.o 00:15:10.399 SO libspdk_vhost.so.8.0 00:15:10.399 CC lib/ftl/nvc/ftl_nvc_dev.o 00:15:10.399 SO libspdk_iscsi.so.8.0 00:15:10.399 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:15:10.399 CC lib/ftl/nvc/ftl_nvc_bdev_non_vss.o 00:15:10.399 SYMLINK libspdk_vhost.so 00:15:10.399 CC lib/ftl/nvc/ftl_nvc_bdev_common.o 00:15:10.399 CC lib/ftl/base/ftl_base_dev.o 00:15:10.399 CC lib/ftl/base/ftl_base_bdev.o 00:15:10.399 CC lib/ftl/ftl_trace.o 00:15:10.399 SYMLINK libspdk_iscsi.so 00:15:10.967 LIB libspdk_ftl.a 00:15:11.225 SO libspdk_ftl.so.9.0 00:15:11.507 SYMLINK libspdk_ftl.so 00:15:12.443 LIB libspdk_nvmf.a 00:15:12.443 SO libspdk_nvmf.so.19.0 00:15:12.701 SYMLINK libspdk_nvmf.so 00:15:12.976 CC module/env_dpdk/env_dpdk_rpc.o 00:15:13.235 CC module/blob/bdev/blob_bdev.o 00:15:13.235 CC module/sock/posix/posix.o 00:15:13.235 CC module/accel/ioat/accel_ioat.o 00:15:13.235 CC module/accel/iaa/accel_iaa.o 00:15:13.235 CC module/scheduler/dynamic/scheduler_dynamic.o 00:15:13.235 CC module/fsdev/aio/fsdev_aio.o 00:15:13.235 CC module/accel/error/accel_error.o 00:15:13.235 CC module/accel/dsa/accel_dsa.o 00:15:13.235 LIB libspdk_env_dpdk_rpc.a 00:15:13.235 CC module/keyring/file/keyring.o 00:15:13.235 SO libspdk_env_dpdk_rpc.so.6.0 00:15:13.235 SYMLINK libspdk_env_dpdk_rpc.so 00:15:13.235 CC module/fsdev/aio/fsdev_aio_rpc.o 00:15:13.492 CC module/keyring/file/keyring_rpc.o 00:15:13.492 CC module/accel/ioat/accel_ioat_rpc.o 00:15:13.492 CC module/accel/error/accel_error_rpc.o 00:15:13.492 LIB libspdk_scheduler_dynamic.a 00:15:13.492 SO libspdk_scheduler_dynamic.so.4.0 00:15:13.492 CC module/accel/iaa/accel_iaa_rpc.o 00:15:13.492 LIB libspdk_blob_bdev.a 00:15:13.492 LIB libspdk_keyring_file.a 00:15:13.492 SO libspdk_blob_bdev.so.11.0 00:15:13.492 SO libspdk_keyring_file.so.2.0 00:15:13.492 SYMLINK libspdk_scheduler_dynamic.so 00:15:13.492 LIB libspdk_accel_ioat.a 00:15:13.492 LIB libspdk_accel_error.a 00:15:13.750 SYMLINK libspdk_blob_bdev.so 00:15:13.750 SO libspdk_accel_ioat.so.6.0 00:15:13.750 SYMLINK libspdk_keyring_file.so 00:15:13.750 SO libspdk_accel_error.so.2.0 00:15:13.750 CC module/accel/dsa/accel_dsa_rpc.o 00:15:13.750 SYMLINK libspdk_accel_ioat.so 00:15:13.750 SYMLINK libspdk_accel_error.so 00:15:13.750 CC module/fsdev/aio/linux_aio_mgr.o 00:15:13.750 LIB libspdk_accel_iaa.a 00:15:13.750 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:15:13.750 SO libspdk_accel_iaa.so.3.0 00:15:13.750 CC module/scheduler/gscheduler/gscheduler.o 00:15:13.750 CC module/keyring/linux/keyring.o 00:15:14.028 LIB libspdk_accel_dsa.a 00:15:14.028 SYMLINK libspdk_accel_iaa.so 00:15:14.028 SO libspdk_accel_dsa.so.5.0 00:15:14.028 CC module/bdev/delay/vbdev_delay.o 00:15:14.028 SYMLINK libspdk_accel_dsa.so 00:15:14.028 CC module/keyring/linux/keyring_rpc.o 00:15:14.028 CC module/blobfs/bdev/blobfs_bdev.o 00:15:14.028 LIB libspdk_scheduler_gscheduler.a 00:15:14.028 LIB libspdk_scheduler_dpdk_governor.a 00:15:14.028 SO libspdk_scheduler_gscheduler.so.4.0 00:15:14.028 LIB libspdk_fsdev_aio.a 00:15:14.299 SO libspdk_scheduler_dpdk_governor.so.4.0 00:15:14.299 CC module/bdev/error/vbdev_error.o 00:15:14.299 SYMLINK libspdk_scheduler_gscheduler.so 00:15:14.299 CC module/bdev/delay/vbdev_delay_rpc.o 00:15:14.299 SO libspdk_fsdev_aio.so.1.0 00:15:14.299 LIB libspdk_sock_posix.a 00:15:14.299 LIB libspdk_keyring_linux.a 00:15:14.299 SO libspdk_sock_posix.so.6.0 00:15:14.299 SYMLINK libspdk_scheduler_dpdk_governor.so 00:15:14.299 SO libspdk_keyring_linux.so.1.0 00:15:14.299 SYMLINK libspdk_fsdev_aio.so 00:15:14.299 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:15:14.299 SYMLINK libspdk_sock_posix.so 00:15:14.299 CC module/bdev/gpt/gpt.o 00:15:14.299 SYMLINK libspdk_keyring_linux.so 00:15:14.299 CC module/bdev/gpt/vbdev_gpt.o 00:15:14.299 CC module/bdev/lvol/vbdev_lvol.o 00:15:14.558 LIB libspdk_bdev_delay.a 00:15:14.558 CC module/bdev/null/bdev_null.o 00:15:14.558 LIB libspdk_blobfs_bdev.a 00:15:14.558 CC module/bdev/malloc/bdev_malloc.o 00:15:14.558 SO libspdk_bdev_delay.so.6.0 00:15:14.558 SO libspdk_blobfs_bdev.so.6.0 00:15:14.558 CC module/bdev/nvme/bdev_nvme.o 00:15:14.558 CC module/bdev/nvme/bdev_nvme_rpc.o 00:15:14.558 CC module/bdev/passthru/vbdev_passthru.o 00:15:14.558 SYMLINK libspdk_bdev_delay.so 00:15:14.558 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:15:14.558 SYMLINK libspdk_blobfs_bdev.so 00:15:14.558 CC module/bdev/nvme/nvme_rpc.o 00:15:14.558 CC module/bdev/error/vbdev_error_rpc.o 00:15:14.816 LIB libspdk_bdev_gpt.a 00:15:14.816 SO libspdk_bdev_gpt.so.6.0 00:15:14.816 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:15:14.816 SYMLINK libspdk_bdev_gpt.so 00:15:14.816 CC module/bdev/null/bdev_null_rpc.o 00:15:15.074 CC module/bdev/nvme/bdev_mdns_client.o 00:15:15.074 LIB libspdk_bdev_error.a 00:15:15.075 SO libspdk_bdev_error.so.6.0 00:15:15.075 LIB libspdk_bdev_passthru.a 00:15:15.075 SYMLINK libspdk_bdev_error.so 00:15:15.075 CC module/bdev/raid/bdev_raid.o 00:15:15.075 CC module/bdev/nvme/vbdev_opal.o 00:15:15.075 SO libspdk_bdev_passthru.so.6.0 00:15:15.075 CC module/bdev/raid/bdev_raid_rpc.o 00:15:15.075 LIB libspdk_bdev_null.a 00:15:15.075 SYMLINK libspdk_bdev_passthru.so 00:15:15.075 SO libspdk_bdev_null.so.6.0 00:15:15.333 CC module/bdev/malloc/bdev_malloc_rpc.o 00:15:15.333 SYMLINK libspdk_bdev_null.so 00:15:15.333 CC module/bdev/split/vbdev_split.o 00:15:15.333 LIB libspdk_bdev_lvol.a 00:15:15.333 CC module/bdev/split/vbdev_split_rpc.o 00:15:15.333 SO libspdk_bdev_lvol.so.6.0 00:15:15.333 CC module/bdev/raid/bdev_raid_sb.o 00:15:15.333 LIB libspdk_bdev_malloc.a 00:15:15.333 CC module/bdev/zone_block/vbdev_zone_block.o 00:15:15.591 SO libspdk_bdev_malloc.so.6.0 00:15:15.591 SYMLINK libspdk_bdev_lvol.so 00:15:15.591 CC module/bdev/raid/raid0.o 00:15:15.591 CC module/bdev/aio/bdev_aio.o 00:15:15.591 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:15:15.591 SYMLINK libspdk_bdev_malloc.so 00:15:15.591 CC module/bdev/aio/bdev_aio_rpc.o 00:15:15.591 LIB libspdk_bdev_split.a 00:15:15.591 SO libspdk_bdev_split.so.6.0 00:15:15.849 CC module/bdev/nvme/vbdev_opal_rpc.o 00:15:15.849 SYMLINK libspdk_bdev_split.so 00:15:15.849 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:15:15.849 CC module/bdev/ftl/bdev_ftl.o 00:15:15.849 CC module/bdev/ftl/bdev_ftl_rpc.o 00:15:15.849 LIB libspdk_bdev_aio.a 00:15:15.849 CC module/bdev/iscsi/bdev_iscsi.o 00:15:16.107 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:15:16.107 SO libspdk_bdev_aio.so.6.0 00:15:16.107 CC module/bdev/raid/raid1.o 00:15:16.107 CC module/bdev/raid/concat.o 00:15:16.107 LIB libspdk_bdev_zone_block.a 00:15:16.107 SYMLINK libspdk_bdev_aio.so 00:15:16.107 CC module/bdev/raid/raid5f.o 00:15:16.107 LIB libspdk_bdev_ftl.a 00:15:16.107 SO libspdk_bdev_zone_block.so.6.0 00:15:16.107 SO libspdk_bdev_ftl.so.6.0 00:15:16.107 SYMLINK libspdk_bdev_zone_block.so 00:15:16.107 SYMLINK libspdk_bdev_ftl.so 00:15:16.366 CC module/bdev/virtio/bdev_virtio_scsi.o 00:15:16.366 CC module/bdev/virtio/bdev_virtio_rpc.o 00:15:16.366 CC module/bdev/virtio/bdev_virtio_blk.o 00:15:16.624 LIB libspdk_bdev_iscsi.a 00:15:16.624 SO libspdk_bdev_iscsi.so.6.0 00:15:16.624 SYMLINK libspdk_bdev_iscsi.so 00:15:16.624 LIB libspdk_bdev_raid.a 00:15:16.882 SO libspdk_bdev_raid.so.6.0 00:15:16.882 LIB libspdk_bdev_virtio.a 00:15:16.882 SYMLINK libspdk_bdev_raid.so 00:15:16.882 SO libspdk_bdev_virtio.so.6.0 00:15:17.140 SYMLINK libspdk_bdev_virtio.so 00:15:18.074 LIB libspdk_bdev_nvme.a 00:15:18.074 SO libspdk_bdev_nvme.so.7.0 00:15:18.074 SYMLINK libspdk_bdev_nvme.so 00:15:18.709 CC module/event/subsystems/keyring/keyring.o 00:15:18.709 CC module/event/subsystems/iobuf/iobuf.o 00:15:18.709 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:15:18.709 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:15:18.710 CC module/event/subsystems/vmd/vmd.o 00:15:18.710 CC module/event/subsystems/vmd/vmd_rpc.o 00:15:18.710 CC module/event/subsystems/sock/sock.o 00:15:18.710 CC module/event/subsystems/scheduler/scheduler.o 00:15:18.710 CC module/event/subsystems/fsdev/fsdev.o 00:15:18.710 LIB libspdk_event_sock.a 00:15:18.710 LIB libspdk_event_iobuf.a 00:15:18.710 LIB libspdk_event_vhost_blk.a 00:15:18.710 LIB libspdk_event_keyring.a 00:15:18.710 LIB libspdk_event_scheduler.a 00:15:18.710 SO libspdk_event_sock.so.5.0 00:15:18.710 SO libspdk_event_iobuf.so.3.0 00:15:18.710 LIB libspdk_event_vmd.a 00:15:18.710 LIB libspdk_event_fsdev.a 00:15:18.710 SO libspdk_event_vhost_blk.so.3.0 00:15:18.710 SO libspdk_event_keyring.so.1.0 00:15:18.710 SO libspdk_event_scheduler.so.4.0 00:15:18.710 SO libspdk_event_fsdev.so.1.0 00:15:18.968 SO libspdk_event_vmd.so.6.0 00:15:18.968 SYMLINK libspdk_event_sock.so 00:15:18.968 SYMLINK libspdk_event_vhost_blk.so 00:15:18.968 SYMLINK libspdk_event_keyring.so 00:15:18.968 SYMLINK libspdk_event_iobuf.so 00:15:18.968 SYMLINK libspdk_event_scheduler.so 00:15:18.968 SYMLINK libspdk_event_fsdev.so 00:15:18.968 SYMLINK libspdk_event_vmd.so 00:15:19.226 CC module/event/subsystems/accel/accel.o 00:15:19.226 LIB libspdk_event_accel.a 00:15:19.483 SO libspdk_event_accel.so.6.0 00:15:19.483 SYMLINK libspdk_event_accel.so 00:15:19.741 CC module/event/subsystems/bdev/bdev.o 00:15:19.998 LIB libspdk_event_bdev.a 00:15:19.998 SO libspdk_event_bdev.so.6.0 00:15:19.998 SYMLINK libspdk_event_bdev.so 00:15:20.255 CC module/event/subsystems/ublk/ublk.o 00:15:20.255 CC module/event/subsystems/scsi/scsi.o 00:15:20.255 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:15:20.255 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:15:20.255 CC module/event/subsystems/nbd/nbd.o 00:15:20.512 LIB libspdk_event_nbd.a 00:15:20.512 LIB libspdk_event_ublk.a 00:15:20.512 SO libspdk_event_nbd.so.6.0 00:15:20.512 SO libspdk_event_ublk.so.3.0 00:15:20.512 LIB libspdk_event_scsi.a 00:15:20.512 SO libspdk_event_scsi.so.6.0 00:15:20.512 SYMLINK libspdk_event_nbd.so 00:15:20.512 SYMLINK libspdk_event_ublk.so 00:15:20.770 SYMLINK libspdk_event_scsi.so 00:15:20.770 LIB libspdk_event_nvmf.a 00:15:20.770 SO libspdk_event_nvmf.so.6.0 00:15:20.770 SYMLINK libspdk_event_nvmf.so 00:15:20.770 CC module/event/subsystems/iscsi/iscsi.o 00:15:20.770 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:15:21.028 LIB libspdk_event_vhost_scsi.a 00:15:21.028 LIB libspdk_event_iscsi.a 00:15:21.028 SO libspdk_event_vhost_scsi.so.3.0 00:15:21.028 SO libspdk_event_iscsi.so.6.0 00:15:21.285 SYMLINK libspdk_event_vhost_scsi.so 00:15:21.285 SYMLINK libspdk_event_iscsi.so 00:15:21.285 SO libspdk.so.6.0 00:15:21.285 SYMLINK libspdk.so 00:15:21.544 CC app/spdk_lspci/spdk_lspci.o 00:15:21.544 CC app/trace_record/trace_record.o 00:15:21.544 CXX app/trace/trace.o 00:15:21.544 CC examples/interrupt_tgt/interrupt_tgt.o 00:15:21.544 CC app/nvmf_tgt/nvmf_main.o 00:15:21.544 CC app/iscsi_tgt/iscsi_tgt.o 00:15:21.802 CC examples/ioat/perf/perf.o 00:15:21.802 CC app/spdk_tgt/spdk_tgt.o 00:15:21.802 CC examples/util/zipf/zipf.o 00:15:21.802 CC test/thread/poller_perf/poller_perf.o 00:15:21.802 LINK spdk_lspci 00:15:21.802 LINK nvmf_tgt 00:15:21.802 LINK interrupt_tgt 00:15:22.061 LINK zipf 00:15:22.061 LINK poller_perf 00:15:22.061 LINK iscsi_tgt 00:15:22.061 LINK spdk_trace_record 00:15:22.061 LINK spdk_tgt 00:15:22.061 LINK ioat_perf 00:15:22.061 CC app/spdk_nvme_perf/perf.o 00:15:22.320 LINK spdk_trace 00:15:22.320 CC app/spdk_nvme_identify/identify.o 00:15:22.320 CC app/spdk_nvme_discover/discovery_aer.o 00:15:22.320 CC app/spdk_top/spdk_top.o 00:15:22.320 CC examples/ioat/verify/verify.o 00:15:22.320 CC test/dma/test_dma/test_dma.o 00:15:22.320 TEST_HEADER include/spdk/accel.h 00:15:22.320 TEST_HEADER include/spdk/accel_module.h 00:15:22.320 TEST_HEADER include/spdk/assert.h 00:15:22.320 TEST_HEADER include/spdk/barrier.h 00:15:22.320 TEST_HEADER include/spdk/base64.h 00:15:22.320 TEST_HEADER include/spdk/bdev.h 00:15:22.320 TEST_HEADER include/spdk/bdev_module.h 00:15:22.320 TEST_HEADER include/spdk/bdev_zone.h 00:15:22.320 TEST_HEADER include/spdk/bit_array.h 00:15:22.320 TEST_HEADER include/spdk/bit_pool.h 00:15:22.320 TEST_HEADER include/spdk/blob_bdev.h 00:15:22.320 TEST_HEADER include/spdk/blobfs_bdev.h 00:15:22.320 TEST_HEADER include/spdk/blobfs.h 00:15:22.320 TEST_HEADER include/spdk/blob.h 00:15:22.320 TEST_HEADER include/spdk/conf.h 00:15:22.320 TEST_HEADER include/spdk/config.h 00:15:22.320 TEST_HEADER include/spdk/cpuset.h 00:15:22.320 TEST_HEADER include/spdk/crc16.h 00:15:22.320 TEST_HEADER include/spdk/crc32.h 00:15:22.320 TEST_HEADER include/spdk/crc64.h 00:15:22.320 TEST_HEADER include/spdk/dif.h 00:15:22.320 TEST_HEADER include/spdk/dma.h 00:15:22.320 TEST_HEADER include/spdk/endian.h 00:15:22.320 TEST_HEADER include/spdk/env_dpdk.h 00:15:22.320 CC test/app/bdev_svc/bdev_svc.o 00:15:22.320 TEST_HEADER include/spdk/env.h 00:15:22.320 TEST_HEADER include/spdk/event.h 00:15:22.320 CC app/spdk_dd/spdk_dd.o 00:15:22.320 TEST_HEADER include/spdk/fd_group.h 00:15:22.320 TEST_HEADER include/spdk/fd.h 00:15:22.320 TEST_HEADER include/spdk/file.h 00:15:22.320 TEST_HEADER include/spdk/fsdev.h 00:15:22.320 TEST_HEADER include/spdk/fsdev_module.h 00:15:22.320 TEST_HEADER include/spdk/ftl.h 00:15:22.320 TEST_HEADER include/spdk/fuse_dispatcher.h 00:15:22.320 TEST_HEADER include/spdk/gpt_spec.h 00:15:22.320 TEST_HEADER include/spdk/hexlify.h 00:15:22.320 TEST_HEADER include/spdk/histogram_data.h 00:15:22.320 TEST_HEADER include/spdk/idxd.h 00:15:22.320 TEST_HEADER include/spdk/idxd_spec.h 00:15:22.320 TEST_HEADER include/spdk/init.h 00:15:22.320 TEST_HEADER include/spdk/ioat.h 00:15:22.320 TEST_HEADER include/spdk/ioat_spec.h 00:15:22.320 TEST_HEADER include/spdk/iscsi_spec.h 00:15:22.320 TEST_HEADER include/spdk/json.h 00:15:22.320 TEST_HEADER include/spdk/jsonrpc.h 00:15:22.320 TEST_HEADER include/spdk/keyring.h 00:15:22.320 TEST_HEADER include/spdk/keyring_module.h 00:15:22.320 TEST_HEADER include/spdk/likely.h 00:15:22.320 TEST_HEADER include/spdk/log.h 00:15:22.320 TEST_HEADER include/spdk/lvol.h 00:15:22.320 TEST_HEADER include/spdk/md5.h 00:15:22.320 TEST_HEADER include/spdk/memory.h 00:15:22.320 TEST_HEADER include/spdk/mmio.h 00:15:22.579 TEST_HEADER include/spdk/nbd.h 00:15:22.579 TEST_HEADER include/spdk/net.h 00:15:22.579 TEST_HEADER include/spdk/notify.h 00:15:22.579 TEST_HEADER include/spdk/nvme.h 00:15:22.579 TEST_HEADER include/spdk/nvme_intel.h 00:15:22.579 TEST_HEADER include/spdk/nvme_ocssd.h 00:15:22.579 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:15:22.579 TEST_HEADER include/spdk/nvme_spec.h 00:15:22.579 TEST_HEADER include/spdk/nvme_zns.h 00:15:22.579 TEST_HEADER include/spdk/nvmf_cmd.h 00:15:22.579 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:15:22.579 TEST_HEADER include/spdk/nvmf.h 00:15:22.579 TEST_HEADER include/spdk/nvmf_spec.h 00:15:22.579 TEST_HEADER include/spdk/nvmf_transport.h 00:15:22.579 TEST_HEADER include/spdk/opal.h 00:15:22.579 TEST_HEADER include/spdk/opal_spec.h 00:15:22.579 TEST_HEADER include/spdk/pci_ids.h 00:15:22.579 TEST_HEADER include/spdk/pipe.h 00:15:22.579 TEST_HEADER include/spdk/queue.h 00:15:22.579 TEST_HEADER include/spdk/reduce.h 00:15:22.579 TEST_HEADER include/spdk/rpc.h 00:15:22.579 TEST_HEADER include/spdk/scheduler.h 00:15:22.579 TEST_HEADER include/spdk/scsi.h 00:15:22.579 LINK spdk_nvme_discover 00:15:22.579 TEST_HEADER include/spdk/scsi_spec.h 00:15:22.579 TEST_HEADER include/spdk/sock.h 00:15:22.579 TEST_HEADER include/spdk/stdinc.h 00:15:22.579 TEST_HEADER include/spdk/string.h 00:15:22.579 TEST_HEADER include/spdk/thread.h 00:15:22.579 TEST_HEADER include/spdk/trace.h 00:15:22.579 TEST_HEADER include/spdk/trace_parser.h 00:15:22.579 TEST_HEADER include/spdk/tree.h 00:15:22.579 TEST_HEADER include/spdk/ublk.h 00:15:22.579 TEST_HEADER include/spdk/util.h 00:15:22.579 TEST_HEADER include/spdk/uuid.h 00:15:22.579 TEST_HEADER include/spdk/version.h 00:15:22.579 TEST_HEADER include/spdk/vfio_user_pci.h 00:15:22.579 TEST_HEADER include/spdk/vfio_user_spec.h 00:15:22.579 TEST_HEADER include/spdk/vhost.h 00:15:22.579 TEST_HEADER include/spdk/vmd.h 00:15:22.579 TEST_HEADER include/spdk/xor.h 00:15:22.579 TEST_HEADER include/spdk/zipf.h 00:15:22.579 CXX test/cpp_headers/accel.o 00:15:22.579 LINK verify 00:15:22.579 LINK bdev_svc 00:15:22.579 CC app/fio/nvme/fio_plugin.o 00:15:22.579 CXX test/cpp_headers/accel_module.o 00:15:22.838 LINK spdk_dd 00:15:22.838 CC app/fio/bdev/fio_plugin.o 00:15:22.838 CXX test/cpp_headers/assert.o 00:15:23.096 LINK test_dma 00:15:23.096 CC examples/thread/thread/thread_ex.o 00:15:23.096 CXX test/cpp_headers/barrier.o 00:15:23.096 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:15:23.096 LINK spdk_nvme_perf 00:15:23.354 CC app/vhost/vhost.o 00:15:23.354 CXX test/cpp_headers/base64.o 00:15:23.354 LINK thread 00:15:23.354 LINK spdk_nvme 00:15:23.354 CXX test/cpp_headers/bdev.o 00:15:23.612 LINK vhost 00:15:23.612 LINK spdk_top 00:15:23.612 CC test/app/histogram_perf/histogram_perf.o 00:15:23.612 LINK spdk_nvme_identify 00:15:23.612 CC examples/sock/hello_world/hello_sock.o 00:15:23.612 LINK spdk_bdev 00:15:23.612 CXX test/cpp_headers/bdev_module.o 00:15:23.612 CXX test/cpp_headers/bdev_zone.o 00:15:23.612 LINK histogram_perf 00:15:23.612 CC examples/vmd/lsvmd/lsvmd.o 00:15:23.612 LINK nvme_fuzz 00:15:23.870 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:15:23.870 LINK lsvmd 00:15:23.870 CC examples/idxd/perf/perf.o 00:15:23.870 CXX test/cpp_headers/bit_array.o 00:15:23.870 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:15:23.870 LINK hello_sock 00:15:23.870 CC examples/accel/perf/accel_perf.o 00:15:23.870 CC examples/fsdev/hello_world/hello_fsdev.o 00:15:23.870 CC test/app/jsoncat/jsoncat.o 00:15:23.870 CC test/app/stub/stub.o 00:15:24.129 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:15:24.129 CXX test/cpp_headers/bit_pool.o 00:15:24.129 LINK jsoncat 00:15:24.129 CC examples/vmd/led/led.o 00:15:24.129 LINK stub 00:15:24.387 LINK hello_fsdev 00:15:24.387 CXX test/cpp_headers/blob_bdev.o 00:15:24.387 CC examples/blob/hello_world/hello_blob.o 00:15:24.387 LINK idxd_perf 00:15:24.387 LINK led 00:15:24.387 CC examples/blob/cli/blobcli.o 00:15:24.387 CXX test/cpp_headers/blobfs_bdev.o 00:15:24.646 CC examples/nvme/hello_world/hello_world.o 00:15:24.646 CC examples/nvme/reconnect/reconnect.o 00:15:24.646 CC examples/nvme/nvme_manage/nvme_manage.o 00:15:24.646 LINK hello_blob 00:15:24.646 LINK vhost_fuzz 00:15:24.646 LINK accel_perf 00:15:24.646 CXX test/cpp_headers/blobfs.o 00:15:24.646 CC test/env/mem_callbacks/mem_callbacks.o 00:15:24.646 CXX test/cpp_headers/blob.o 00:15:24.905 CXX test/cpp_headers/conf.o 00:15:24.905 CXX test/cpp_headers/config.o 00:15:24.905 LINK hello_world 00:15:24.905 CXX test/cpp_headers/cpuset.o 00:15:24.905 CC test/env/vtophys/vtophys.o 00:15:24.905 LINK reconnect 00:15:25.163 LINK blobcli 00:15:25.163 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:15:25.163 CXX test/cpp_headers/crc16.o 00:15:25.163 CC examples/nvme/arbitration/arbitration.o 00:15:25.163 CC test/env/memory/memory_ut.o 00:15:25.163 LINK vtophys 00:15:25.163 LINK nvme_manage 00:15:25.163 CC test/env/pci/pci_ut.o 00:15:25.421 LINK env_dpdk_post_init 00:15:25.421 CXX test/cpp_headers/crc32.o 00:15:25.421 CXX test/cpp_headers/crc64.o 00:15:25.421 LINK mem_callbacks 00:15:25.421 CC examples/bdev/hello_world/hello_bdev.o 00:15:25.421 CC examples/nvme/hotplug/hotplug.o 00:15:25.421 CXX test/cpp_headers/dif.o 00:15:25.678 LINK arbitration 00:15:25.678 CC examples/nvme/abort/abort.o 00:15:25.678 CC examples/nvme/cmb_copy/cmb_copy.o 00:15:25.678 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:15:25.678 CXX test/cpp_headers/dma.o 00:15:25.678 CXX test/cpp_headers/endian.o 00:15:25.678 LINK pci_ut 00:15:25.678 LINK hello_bdev 00:15:25.937 LINK hotplug 00:15:25.937 LINK cmb_copy 00:15:25.937 LINK pmr_persistence 00:15:25.937 CXX test/cpp_headers/env_dpdk.o 00:15:25.937 CC examples/bdev/bdevperf/bdevperf.o 00:15:25.937 LINK iscsi_fuzz 00:15:26.199 CXX test/cpp_headers/env.o 00:15:26.199 LINK abort 00:15:26.199 CC test/rpc_client/rpc_client_test.o 00:15:26.199 CC test/event/event_perf/event_perf.o 00:15:26.199 CC test/nvme/aer/aer.o 00:15:26.199 CC test/event/reactor/reactor.o 00:15:26.199 CXX test/cpp_headers/event.o 00:15:26.199 CXX test/cpp_headers/fd_group.o 00:15:26.199 CC test/accel/dif/dif.o 00:15:26.199 CXX test/cpp_headers/fd.o 00:15:26.485 LINK rpc_client_test 00:15:26.485 LINK event_perf 00:15:26.485 LINK reactor 00:15:26.485 CXX test/cpp_headers/file.o 00:15:26.485 CXX test/cpp_headers/fsdev.o 00:15:26.485 CC test/nvme/reset/reset.o 00:15:26.485 CC test/nvme/sgl/sgl.o 00:15:26.743 LINK aer 00:15:26.743 LINK memory_ut 00:15:26.743 CC test/event/reactor_perf/reactor_perf.o 00:15:26.743 CC test/event/app_repeat/app_repeat.o 00:15:26.743 CXX test/cpp_headers/fsdev_module.o 00:15:26.743 CC test/event/scheduler/scheduler.o 00:15:26.743 LINK reactor_perf 00:15:27.000 LINK app_repeat 00:15:27.000 CC test/nvme/e2edp/nvme_dp.o 00:15:27.000 LINK reset 00:15:27.000 CXX test/cpp_headers/ftl.o 00:15:27.000 LINK sgl 00:15:27.000 CC test/nvme/overhead/overhead.o 00:15:27.000 CXX test/cpp_headers/fuse_dispatcher.o 00:15:27.000 LINK bdevperf 00:15:27.000 CC test/nvme/err_injection/err_injection.o 00:15:27.000 LINK scheduler 00:15:27.258 CC test/nvme/startup/startup.o 00:15:27.258 LINK dif 00:15:27.258 CC test/nvme/reserve/reserve.o 00:15:27.258 LINK nvme_dp 00:15:27.258 CXX test/cpp_headers/gpt_spec.o 00:15:27.258 CC test/nvme/simple_copy/simple_copy.o 00:15:27.258 LINK err_injection 00:15:27.258 LINK overhead 00:15:27.517 LINK startup 00:15:27.517 CXX test/cpp_headers/hexlify.o 00:15:27.517 CC examples/nvmf/nvmf/nvmf.o 00:15:27.517 LINK reserve 00:15:27.517 LINK simple_copy 00:15:27.775 CC test/nvme/connect_stress/connect_stress.o 00:15:27.775 CC test/blobfs/mkfs/mkfs.o 00:15:27.775 CC test/nvme/boot_partition/boot_partition.o 00:15:27.775 CXX test/cpp_headers/histogram_data.o 00:15:27.775 CC test/nvme/compliance/nvme_compliance.o 00:15:27.775 CC test/bdev/bdevio/bdevio.o 00:15:27.775 CC test/lvol/esnap/esnap.o 00:15:27.775 LINK connect_stress 00:15:27.775 CC test/nvme/fused_ordering/fused_ordering.o 00:15:27.775 CC test/nvme/doorbell_aers/doorbell_aers.o 00:15:27.775 LINK mkfs 00:15:27.775 LINK boot_partition 00:15:28.033 CXX test/cpp_headers/idxd.o 00:15:28.033 LINK nvmf 00:15:28.033 LINK doorbell_aers 00:15:28.033 CXX test/cpp_headers/idxd_spec.o 00:15:28.033 LINK fused_ordering 00:15:28.033 CXX test/cpp_headers/init.o 00:15:28.033 LINK nvme_compliance 00:15:28.291 CC test/nvme/cuse/cuse.o 00:15:28.291 CC test/nvme/fdp/fdp.o 00:15:28.291 CXX test/cpp_headers/ioat.o 00:15:28.291 LINK bdevio 00:15:28.291 CXX test/cpp_headers/ioat_spec.o 00:15:28.291 CXX test/cpp_headers/iscsi_spec.o 00:15:28.291 CXX test/cpp_headers/json.o 00:15:28.291 CXX test/cpp_headers/jsonrpc.o 00:15:28.291 CXX test/cpp_headers/keyring.o 00:15:28.291 CXX test/cpp_headers/keyring_module.o 00:15:28.549 CXX test/cpp_headers/likely.o 00:15:28.549 CXX test/cpp_headers/log.o 00:15:28.549 CXX test/cpp_headers/lvol.o 00:15:28.549 CXX test/cpp_headers/md5.o 00:15:28.549 CXX test/cpp_headers/memory.o 00:15:28.549 CXX test/cpp_headers/mmio.o 00:15:28.549 CXX test/cpp_headers/nbd.o 00:15:28.549 CXX test/cpp_headers/net.o 00:15:28.549 LINK fdp 00:15:28.808 CXX test/cpp_headers/notify.o 00:15:28.808 CXX test/cpp_headers/nvme.o 00:15:28.808 CXX test/cpp_headers/nvme_intel.o 00:15:28.808 CXX test/cpp_headers/nvme_ocssd.o 00:15:28.808 CXX test/cpp_headers/nvme_ocssd_spec.o 00:15:28.808 CXX test/cpp_headers/nvme_spec.o 00:15:28.808 CXX test/cpp_headers/nvme_zns.o 00:15:28.808 CXX test/cpp_headers/nvmf_cmd.o 00:15:28.808 CXX test/cpp_headers/nvmf_fc_spec.o 00:15:28.808 CXX test/cpp_headers/nvmf.o 00:15:28.808 CXX test/cpp_headers/nvmf_spec.o 00:15:28.808 CXX test/cpp_headers/nvmf_transport.o 00:15:28.808 CXX test/cpp_headers/opal.o 00:15:29.069 CXX test/cpp_headers/opal_spec.o 00:15:29.069 CXX test/cpp_headers/pci_ids.o 00:15:29.069 CXX test/cpp_headers/pipe.o 00:15:29.069 CXX test/cpp_headers/queue.o 00:15:29.069 CXX test/cpp_headers/reduce.o 00:15:29.069 CXX test/cpp_headers/rpc.o 00:15:29.069 CXX test/cpp_headers/scheduler.o 00:15:29.069 CXX test/cpp_headers/scsi.o 00:15:29.069 CXX test/cpp_headers/scsi_spec.o 00:15:29.069 CXX test/cpp_headers/sock.o 00:15:29.069 CXX test/cpp_headers/stdinc.o 00:15:29.326 CXX test/cpp_headers/string.o 00:15:29.326 CXX test/cpp_headers/thread.o 00:15:29.326 CXX test/cpp_headers/trace.o 00:15:29.326 CXX test/cpp_headers/trace_parser.o 00:15:29.326 CXX test/cpp_headers/tree.o 00:15:29.326 CXX test/cpp_headers/ublk.o 00:15:29.326 CXX test/cpp_headers/util.o 00:15:29.326 CXX test/cpp_headers/uuid.o 00:15:29.326 CXX test/cpp_headers/version.o 00:15:29.326 CXX test/cpp_headers/vfio_user_pci.o 00:15:29.326 CXX test/cpp_headers/vfio_user_spec.o 00:15:29.584 CXX test/cpp_headers/vhost.o 00:15:29.584 CXX test/cpp_headers/vmd.o 00:15:29.584 CXX test/cpp_headers/xor.o 00:15:29.584 CXX test/cpp_headers/zipf.o 00:15:29.842 LINK cuse 00:15:35.107 LINK esnap 00:15:35.366 00:15:35.366 real 1m45.653s 00:15:35.366 user 9m26.533s 00:15:35.366 sys 1m53.428s 00:15:35.366 20:14:30 make -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:15:35.366 20:14:30 make -- common/autotest_common.sh@10 -- $ set +x 00:15:35.366 ************************************ 00:15:35.366 END TEST make 00:15:35.366 ************************************ 00:15:35.366 20:14:30 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:15:35.366 20:14:30 -- pm/common@29 -- $ signal_monitor_resources TERM 00:15:35.366 20:14:30 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:15:35.366 20:14:30 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:15:35.366 20:14:30 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-cpu-load.pid ]] 00:15:35.366 20:14:30 -- pm/common@44 -- $ pid=5238 00:15:35.366 20:14:30 -- pm/common@50 -- $ kill -TERM 5238 00:15:35.366 20:14:30 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:15:35.366 20:14:30 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-vmstat.pid ]] 00:15:35.366 20:14:30 -- pm/common@44 -- $ pid=5240 00:15:35.366 20:14:30 -- pm/common@50 -- $ kill -TERM 5240 00:15:35.624 20:14:30 -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:15:35.625 20:14:30 -- common/autotest_common.sh@1681 -- # lcov --version 00:15:35.625 20:14:30 -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:15:35.625 20:14:30 -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:15:35.625 20:14:30 -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:15:35.625 20:14:30 -- scripts/common.sh@333 -- # local ver1 ver1_l 00:15:35.625 20:14:30 -- scripts/common.sh@334 -- # local ver2 ver2_l 00:15:35.625 20:14:30 -- scripts/common.sh@336 -- # IFS=.-: 00:15:35.625 20:14:30 -- scripts/common.sh@336 -- # read -ra ver1 00:15:35.625 20:14:30 -- scripts/common.sh@337 -- # IFS=.-: 00:15:35.625 20:14:30 -- scripts/common.sh@337 -- # read -ra ver2 00:15:35.625 20:14:30 -- scripts/common.sh@338 -- # local 'op=<' 00:15:35.625 20:14:30 -- scripts/common.sh@340 -- # ver1_l=2 00:15:35.625 20:14:30 -- scripts/common.sh@341 -- # ver2_l=1 00:15:35.625 20:14:30 -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:15:35.625 20:14:30 -- scripts/common.sh@344 -- # case "$op" in 00:15:35.625 20:14:30 -- scripts/common.sh@345 -- # : 1 00:15:35.625 20:14:30 -- scripts/common.sh@364 -- # (( v = 0 )) 00:15:35.625 20:14:30 -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:15:35.625 20:14:30 -- scripts/common.sh@365 -- # decimal 1 00:15:35.625 20:14:30 -- scripts/common.sh@353 -- # local d=1 00:15:35.625 20:14:30 -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:15:35.625 20:14:30 -- scripts/common.sh@355 -- # echo 1 00:15:35.625 20:14:30 -- scripts/common.sh@365 -- # ver1[v]=1 00:15:35.625 20:14:30 -- scripts/common.sh@366 -- # decimal 2 00:15:35.625 20:14:30 -- scripts/common.sh@353 -- # local d=2 00:15:35.625 20:14:30 -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:15:35.625 20:14:30 -- scripts/common.sh@355 -- # echo 2 00:15:35.625 20:14:30 -- scripts/common.sh@366 -- # ver2[v]=2 00:15:35.625 20:14:30 -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:15:35.625 20:14:30 -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:15:35.625 20:14:30 -- scripts/common.sh@368 -- # return 0 00:15:35.625 20:14:30 -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:15:35.625 20:14:30 -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:15:35.625 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:35.625 --rc genhtml_branch_coverage=1 00:15:35.625 --rc genhtml_function_coverage=1 00:15:35.625 --rc genhtml_legend=1 00:15:35.625 --rc geninfo_all_blocks=1 00:15:35.625 --rc geninfo_unexecuted_blocks=1 00:15:35.625 00:15:35.625 ' 00:15:35.625 20:14:30 -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:15:35.625 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:35.625 --rc genhtml_branch_coverage=1 00:15:35.625 --rc genhtml_function_coverage=1 00:15:35.625 --rc genhtml_legend=1 00:15:35.625 --rc geninfo_all_blocks=1 00:15:35.625 --rc geninfo_unexecuted_blocks=1 00:15:35.625 00:15:35.625 ' 00:15:35.625 20:14:30 -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:15:35.625 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:35.625 --rc genhtml_branch_coverage=1 00:15:35.625 --rc genhtml_function_coverage=1 00:15:35.625 --rc genhtml_legend=1 00:15:35.625 --rc geninfo_all_blocks=1 00:15:35.625 --rc geninfo_unexecuted_blocks=1 00:15:35.625 00:15:35.625 ' 00:15:35.625 20:14:30 -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:15:35.625 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:35.625 --rc genhtml_branch_coverage=1 00:15:35.625 --rc genhtml_function_coverage=1 00:15:35.625 --rc genhtml_legend=1 00:15:35.625 --rc geninfo_all_blocks=1 00:15:35.625 --rc geninfo_unexecuted_blocks=1 00:15:35.625 00:15:35.625 ' 00:15:35.625 20:14:30 -- spdk/autotest.sh@25 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:15:35.625 20:14:30 -- nvmf/common.sh@7 -- # uname -s 00:15:35.625 20:14:30 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:35.625 20:14:30 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:35.625 20:14:30 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:35.625 20:14:30 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:35.625 20:14:30 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:35.625 20:14:30 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:35.625 20:14:30 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:35.625 20:14:30 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:35.625 20:14:30 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:35.625 20:14:30 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:35.625 20:14:30 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5a683382-c549-400e-8b27-e29f159572f4 00:15:35.625 20:14:30 -- nvmf/common.sh@18 -- # NVME_HOSTID=5a683382-c549-400e-8b27-e29f159572f4 00:15:35.625 20:14:30 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:35.625 20:14:30 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:35.625 20:14:30 -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:15:35.625 20:14:30 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:35.625 20:14:30 -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:15:35.625 20:14:30 -- scripts/common.sh@15 -- # shopt -s extglob 00:15:35.625 20:14:30 -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:35.625 20:14:30 -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:35.625 20:14:30 -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:35.625 20:14:30 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:35.625 20:14:30 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:35.625 20:14:30 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:35.625 20:14:30 -- paths/export.sh@5 -- # export PATH 00:15:35.625 20:14:30 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:35.625 20:14:30 -- nvmf/common.sh@51 -- # : 0 00:15:35.625 20:14:30 -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:15:35.625 20:14:30 -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:15:35.625 20:14:30 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:35.625 20:14:30 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:35.625 20:14:30 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:35.625 20:14:30 -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:15:35.625 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:15:35.625 20:14:30 -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:15:35.625 20:14:30 -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:15:35.625 20:14:30 -- nvmf/common.sh@55 -- # have_pci_nics=0 00:15:35.625 20:14:30 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:15:35.625 20:14:30 -- spdk/autotest.sh@32 -- # uname -s 00:15:35.625 20:14:30 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:15:35.625 20:14:30 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:15:35.625 20:14:30 -- spdk/autotest.sh@34 -- # mkdir -p /home/vagrant/spdk_repo/spdk/../output/coredumps 00:15:35.625 20:14:30 -- spdk/autotest.sh@39 -- # echo '|/home/vagrant/spdk_repo/spdk/scripts/core-collector.sh %P %s %t' 00:15:35.625 20:14:30 -- spdk/autotest.sh@40 -- # echo /home/vagrant/spdk_repo/spdk/../output/coredumps 00:15:35.625 20:14:30 -- spdk/autotest.sh@44 -- # modprobe nbd 00:15:35.884 20:14:30 -- spdk/autotest.sh@46 -- # type -P udevadm 00:15:35.884 20:14:30 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:15:35.884 20:14:30 -- spdk/autotest.sh@48 -- # udevadm_pid=54391 00:15:35.884 20:14:30 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:15:35.884 20:14:30 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:15:35.884 20:14:30 -- pm/common@17 -- # local monitor 00:15:35.884 20:14:30 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:15:35.884 20:14:30 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:15:35.884 20:14:30 -- pm/common@25 -- # sleep 1 00:15:35.884 20:14:30 -- pm/common@21 -- # date +%s 00:15:35.884 20:14:30 -- pm/common@21 -- # date +%s 00:15:35.885 20:14:30 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1727813670 00:15:35.885 20:14:30 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1727813670 00:15:35.885 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1727813670_collect-cpu-load.pm.log 00:15:35.885 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1727813670_collect-vmstat.pm.log 00:15:36.821 20:14:31 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:15:36.821 20:14:31 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:15:36.821 20:14:31 -- common/autotest_common.sh@724 -- # xtrace_disable 00:15:36.821 20:14:31 -- common/autotest_common.sh@10 -- # set +x 00:15:36.821 20:14:31 -- spdk/autotest.sh@59 -- # create_test_list 00:15:36.821 20:14:31 -- common/autotest_common.sh@748 -- # xtrace_disable 00:15:36.821 20:14:31 -- common/autotest_common.sh@10 -- # set +x 00:15:36.821 20:14:31 -- spdk/autotest.sh@61 -- # dirname /home/vagrant/spdk_repo/spdk/autotest.sh 00:15:36.821 20:14:31 -- spdk/autotest.sh@61 -- # readlink -f /home/vagrant/spdk_repo/spdk 00:15:36.821 20:14:31 -- spdk/autotest.sh@61 -- # src=/home/vagrant/spdk_repo/spdk 00:15:36.821 20:14:31 -- spdk/autotest.sh@62 -- # out=/home/vagrant/spdk_repo/spdk/../output 00:15:36.821 20:14:31 -- spdk/autotest.sh@63 -- # cd /home/vagrant/spdk_repo/spdk 00:15:36.821 20:14:31 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:15:36.821 20:14:31 -- common/autotest_common.sh@1455 -- # uname 00:15:36.821 20:14:31 -- common/autotest_common.sh@1455 -- # '[' Linux = FreeBSD ']' 00:15:36.821 20:14:31 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:15:36.821 20:14:31 -- common/autotest_common.sh@1475 -- # uname 00:15:36.821 20:14:31 -- common/autotest_common.sh@1475 -- # [[ Linux = FreeBSD ]] 00:15:36.821 20:14:31 -- spdk/autotest.sh@68 -- # [[ y == y ]] 00:15:36.821 20:14:31 -- spdk/autotest.sh@70 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 --version 00:15:36.821 lcov: LCOV version 1.15 00:15:36.821 20:14:32 -- spdk/autotest.sh@72 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -i -t Baseline -d /home/vagrant/spdk_repo/spdk -o /home/vagrant/spdk_repo/spdk/../output/cov_base.info 00:15:54.967 /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:15:54.967 geninfo: WARNING: GCOV did not produce any data for /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno 00:16:13.050 20:15:05 -- spdk/autotest.sh@76 -- # timing_enter pre_cleanup 00:16:13.050 20:15:05 -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:13.050 20:15:05 -- common/autotest_common.sh@10 -- # set +x 00:16:13.050 20:15:05 -- spdk/autotest.sh@78 -- # rm -f 00:16:13.050 20:15:05 -- spdk/autotest.sh@81 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:16:13.050 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:16:13.050 0000:00:11.0 (1b36 0010): Already using the nvme driver 00:16:13.050 0000:00:10.0 (1b36 0010): Already using the nvme driver 00:16:13.050 20:15:06 -- spdk/autotest.sh@83 -- # get_zoned_devs 00:16:13.050 20:15:06 -- common/autotest_common.sh@1655 -- # zoned_devs=() 00:16:13.050 20:15:06 -- common/autotest_common.sh@1655 -- # local -gA zoned_devs 00:16:13.050 20:15:06 -- common/autotest_common.sh@1656 -- # local nvme bdf 00:16:13.050 20:15:06 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:16:13.050 20:15:06 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme0n1 00:16:13.050 20:15:06 -- common/autotest_common.sh@1648 -- # local device=nvme0n1 00:16:13.050 20:15:06 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:16:13.050 20:15:06 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:16:13.050 20:15:06 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:16:13.050 20:15:06 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme1n1 00:16:13.050 20:15:06 -- common/autotest_common.sh@1648 -- # local device=nvme1n1 00:16:13.050 20:15:06 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme1n1/queue/zoned ]] 00:16:13.050 20:15:06 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:16:13.050 20:15:06 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:16:13.050 20:15:06 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme1n2 00:16:13.050 20:15:06 -- common/autotest_common.sh@1648 -- # local device=nvme1n2 00:16:13.050 20:15:06 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme1n2/queue/zoned ]] 00:16:13.050 20:15:06 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:16:13.050 20:15:06 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:16:13.050 20:15:06 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme1n3 00:16:13.050 20:15:06 -- common/autotest_common.sh@1648 -- # local device=nvme1n3 00:16:13.050 20:15:06 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme1n3/queue/zoned ]] 00:16:13.050 20:15:06 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:16:13.050 20:15:06 -- spdk/autotest.sh@85 -- # (( 0 > 0 )) 00:16:13.050 20:15:06 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:16:13.050 20:15:06 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:16:13.050 20:15:06 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n1 00:16:13.050 20:15:06 -- scripts/common.sh@381 -- # local block=/dev/nvme0n1 pt 00:16:13.050 20:15:06 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:16:13.050 No valid GPT data, bailing 00:16:13.050 20:15:06 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:16:13.050 20:15:06 -- scripts/common.sh@394 -- # pt= 00:16:13.050 20:15:06 -- scripts/common.sh@395 -- # return 1 00:16:13.051 20:15:06 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:16:13.051 1+0 records in 00:16:13.051 1+0 records out 00:16:13.051 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00502839 s, 209 MB/s 00:16:13.051 20:15:06 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:16:13.051 20:15:06 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:16:13.051 20:15:06 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n1 00:16:13.051 20:15:06 -- scripts/common.sh@381 -- # local block=/dev/nvme1n1 pt 00:16:13.051 20:15:06 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n1 00:16:13.051 No valid GPT data, bailing 00:16:13.051 20:15:06 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n1 00:16:13.051 20:15:06 -- scripts/common.sh@394 -- # pt= 00:16:13.051 20:15:06 -- scripts/common.sh@395 -- # return 1 00:16:13.051 20:15:06 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n1 bs=1M count=1 00:16:13.051 1+0 records in 00:16:13.051 1+0 records out 00:16:13.051 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00410355 s, 256 MB/s 00:16:13.051 20:15:06 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:16:13.051 20:15:06 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:16:13.051 20:15:06 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n2 00:16:13.051 20:15:06 -- scripts/common.sh@381 -- # local block=/dev/nvme1n2 pt 00:16:13.051 20:15:06 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n2 00:16:13.051 No valid GPT data, bailing 00:16:13.051 20:15:06 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n2 00:16:13.051 20:15:06 -- scripts/common.sh@394 -- # pt= 00:16:13.051 20:15:06 -- scripts/common.sh@395 -- # return 1 00:16:13.051 20:15:06 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n2 bs=1M count=1 00:16:13.051 1+0 records in 00:16:13.051 1+0 records out 00:16:13.051 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00398232 s, 263 MB/s 00:16:13.051 20:15:06 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:16:13.051 20:15:06 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:16:13.051 20:15:06 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n3 00:16:13.051 20:15:06 -- scripts/common.sh@381 -- # local block=/dev/nvme1n3 pt 00:16:13.051 20:15:06 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n3 00:16:13.051 No valid GPT data, bailing 00:16:13.051 20:15:06 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n3 00:16:13.051 20:15:06 -- scripts/common.sh@394 -- # pt= 00:16:13.051 20:15:06 -- scripts/common.sh@395 -- # return 1 00:16:13.051 20:15:06 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n3 bs=1M count=1 00:16:13.051 1+0 records in 00:16:13.051 1+0 records out 00:16:13.051 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00400128 s, 262 MB/s 00:16:13.051 20:15:06 -- spdk/autotest.sh@105 -- # sync 00:16:13.051 20:15:06 -- spdk/autotest.sh@107 -- # xtrace_disable_per_cmd reap_spdk_processes 00:16:13.051 20:15:06 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:16:13.051 20:15:06 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:16:13.618 20:15:08 -- spdk/autotest.sh@111 -- # uname -s 00:16:13.618 20:15:08 -- spdk/autotest.sh@111 -- # [[ Linux == Linux ]] 00:16:13.618 20:15:08 -- spdk/autotest.sh@111 -- # [[ 0 -eq 1 ]] 00:16:13.618 20:15:08 -- spdk/autotest.sh@115 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:16:14.186 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:16:14.186 Hugepages 00:16:14.186 node hugesize free / total 00:16:14.186 node0 1048576kB 0 / 0 00:16:14.186 node0 2048kB 0 / 0 00:16:14.186 00:16:14.186 Type BDF Vendor Device NUMA Driver Device Block devices 00:16:14.186 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:16:14.444 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:16:14.444 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:16:14.444 20:15:09 -- spdk/autotest.sh@117 -- # uname -s 00:16:14.444 20:15:09 -- spdk/autotest.sh@117 -- # [[ Linux == Linux ]] 00:16:14.444 20:15:09 -- spdk/autotest.sh@119 -- # nvme_namespace_revert 00:16:14.444 20:15:09 -- common/autotest_common.sh@1514 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:16:15.009 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:16:15.268 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:16:15.268 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:16:15.268 20:15:10 -- common/autotest_common.sh@1515 -- # sleep 1 00:16:16.202 20:15:11 -- common/autotest_common.sh@1516 -- # bdfs=() 00:16:16.202 20:15:11 -- common/autotest_common.sh@1516 -- # local bdfs 00:16:16.202 20:15:11 -- common/autotest_common.sh@1518 -- # bdfs=($(get_nvme_bdfs)) 00:16:16.202 20:15:11 -- common/autotest_common.sh@1518 -- # get_nvme_bdfs 00:16:16.202 20:15:11 -- common/autotest_common.sh@1496 -- # bdfs=() 00:16:16.202 20:15:11 -- common/autotest_common.sh@1496 -- # local bdfs 00:16:16.202 20:15:11 -- common/autotest_common.sh@1497 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:16:16.202 20:15:11 -- common/autotest_common.sh@1497 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:16:16.202 20:15:11 -- common/autotest_common.sh@1497 -- # jq -r '.config[].params.traddr' 00:16:16.460 20:15:11 -- common/autotest_common.sh@1498 -- # (( 2 == 0 )) 00:16:16.460 20:15:11 -- common/autotest_common.sh@1502 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:16:16.460 20:15:11 -- common/autotest_common.sh@1520 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:16:16.718 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:16:16.718 Waiting for block devices as requested 00:16:16.718 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:16:16.977 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:16:16.977 20:15:12 -- common/autotest_common.sh@1522 -- # for bdf in "${bdfs[@]}" 00:16:16.977 20:15:12 -- common/autotest_common.sh@1523 -- # get_nvme_ctrlr_from_bdf 0000:00:10.0 00:16:16.977 20:15:12 -- common/autotest_common.sh@1485 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:16:16.977 20:15:12 -- common/autotest_common.sh@1485 -- # grep 0000:00:10.0/nvme/nvme 00:16:16.977 20:15:12 -- common/autotest_common.sh@1485 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:16:16.977 20:15:12 -- common/autotest_common.sh@1486 -- # [[ -z /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 ]] 00:16:16.977 20:15:12 -- common/autotest_common.sh@1490 -- # basename /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:16:16.977 20:15:12 -- common/autotest_common.sh@1490 -- # printf '%s\n' nvme1 00:16:16.977 20:15:12 -- common/autotest_common.sh@1523 -- # nvme_ctrlr=/dev/nvme1 00:16:16.977 20:15:12 -- common/autotest_common.sh@1524 -- # [[ -z /dev/nvme1 ]] 00:16:16.977 20:15:12 -- common/autotest_common.sh@1529 -- # nvme id-ctrl /dev/nvme1 00:16:16.977 20:15:12 -- common/autotest_common.sh@1529 -- # grep oacs 00:16:16.977 20:15:12 -- common/autotest_common.sh@1529 -- # cut -d: -f2 00:16:16.977 20:15:12 -- common/autotest_common.sh@1529 -- # oacs=' 0x12a' 00:16:16.977 20:15:12 -- common/autotest_common.sh@1530 -- # oacs_ns_manage=8 00:16:16.977 20:15:12 -- common/autotest_common.sh@1532 -- # [[ 8 -ne 0 ]] 00:16:16.977 20:15:12 -- common/autotest_common.sh@1538 -- # nvme id-ctrl /dev/nvme1 00:16:16.977 20:15:12 -- common/autotest_common.sh@1538 -- # grep unvmcap 00:16:16.977 20:15:12 -- common/autotest_common.sh@1538 -- # cut -d: -f2 00:16:16.977 20:15:12 -- common/autotest_common.sh@1538 -- # unvmcap=' 0' 00:16:16.977 20:15:12 -- common/autotest_common.sh@1539 -- # [[ 0 -eq 0 ]] 00:16:16.977 20:15:12 -- common/autotest_common.sh@1541 -- # continue 00:16:16.977 20:15:12 -- common/autotest_common.sh@1522 -- # for bdf in "${bdfs[@]}" 00:16:16.977 20:15:12 -- common/autotest_common.sh@1523 -- # get_nvme_ctrlr_from_bdf 0000:00:11.0 00:16:16.977 20:15:12 -- common/autotest_common.sh@1485 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:16:16.977 20:15:12 -- common/autotest_common.sh@1485 -- # grep 0000:00:11.0/nvme/nvme 00:16:16.977 20:15:12 -- common/autotest_common.sh@1485 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:16:16.977 20:15:12 -- common/autotest_common.sh@1486 -- # [[ -z /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 ]] 00:16:16.977 20:15:12 -- common/autotest_common.sh@1490 -- # basename /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:16:16.977 20:15:12 -- common/autotest_common.sh@1490 -- # printf '%s\n' nvme0 00:16:16.977 20:15:12 -- common/autotest_common.sh@1523 -- # nvme_ctrlr=/dev/nvme0 00:16:16.977 20:15:12 -- common/autotest_common.sh@1524 -- # [[ -z /dev/nvme0 ]] 00:16:16.977 20:15:12 -- common/autotest_common.sh@1529 -- # nvme id-ctrl /dev/nvme0 00:16:16.977 20:15:12 -- common/autotest_common.sh@1529 -- # cut -d: -f2 00:16:16.977 20:15:12 -- common/autotest_common.sh@1529 -- # grep oacs 00:16:16.977 20:15:12 -- common/autotest_common.sh@1529 -- # oacs=' 0x12a' 00:16:16.977 20:15:12 -- common/autotest_common.sh@1530 -- # oacs_ns_manage=8 00:16:16.977 20:15:12 -- common/autotest_common.sh@1532 -- # [[ 8 -ne 0 ]] 00:16:16.977 20:15:12 -- common/autotest_common.sh@1538 -- # nvme id-ctrl /dev/nvme0 00:16:16.977 20:15:12 -- common/autotest_common.sh@1538 -- # grep unvmcap 00:16:16.977 20:15:12 -- common/autotest_common.sh@1538 -- # cut -d: -f2 00:16:16.977 20:15:12 -- common/autotest_common.sh@1538 -- # unvmcap=' 0' 00:16:16.977 20:15:12 -- common/autotest_common.sh@1539 -- # [[ 0 -eq 0 ]] 00:16:16.977 20:15:12 -- common/autotest_common.sh@1541 -- # continue 00:16:16.977 20:15:12 -- spdk/autotest.sh@122 -- # timing_exit pre_cleanup 00:16:16.977 20:15:12 -- common/autotest_common.sh@730 -- # xtrace_disable 00:16:16.977 20:15:12 -- common/autotest_common.sh@10 -- # set +x 00:16:16.977 20:15:12 -- spdk/autotest.sh@125 -- # timing_enter afterboot 00:16:16.978 20:15:12 -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:16.978 20:15:12 -- common/autotest_common.sh@10 -- # set +x 00:16:17.235 20:15:12 -- spdk/autotest.sh@126 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:16:17.800 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:16:17.800 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:16:18.059 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:16:18.059 20:15:13 -- spdk/autotest.sh@127 -- # timing_exit afterboot 00:16:18.059 20:15:13 -- common/autotest_common.sh@730 -- # xtrace_disable 00:16:18.059 20:15:13 -- common/autotest_common.sh@10 -- # set +x 00:16:18.059 20:15:13 -- spdk/autotest.sh@131 -- # opal_revert_cleanup 00:16:18.059 20:15:13 -- common/autotest_common.sh@1576 -- # mapfile -t bdfs 00:16:18.059 20:15:13 -- common/autotest_common.sh@1576 -- # get_nvme_bdfs_by_id 0x0a54 00:16:18.059 20:15:13 -- common/autotest_common.sh@1561 -- # bdfs=() 00:16:18.059 20:15:13 -- common/autotest_common.sh@1561 -- # _bdfs=() 00:16:18.059 20:15:13 -- common/autotest_common.sh@1561 -- # local bdfs _bdfs 00:16:18.059 20:15:13 -- common/autotest_common.sh@1562 -- # _bdfs=($(get_nvme_bdfs)) 00:16:18.059 20:15:13 -- common/autotest_common.sh@1562 -- # get_nvme_bdfs 00:16:18.059 20:15:13 -- common/autotest_common.sh@1496 -- # bdfs=() 00:16:18.059 20:15:13 -- common/autotest_common.sh@1496 -- # local bdfs 00:16:18.059 20:15:13 -- common/autotest_common.sh@1497 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:16:18.059 20:15:13 -- common/autotest_common.sh@1497 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:16:18.059 20:15:13 -- common/autotest_common.sh@1497 -- # jq -r '.config[].params.traddr' 00:16:18.059 20:15:13 -- common/autotest_common.sh@1498 -- # (( 2 == 0 )) 00:16:18.059 20:15:13 -- common/autotest_common.sh@1502 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:16:18.059 20:15:13 -- common/autotest_common.sh@1563 -- # for bdf in "${_bdfs[@]}" 00:16:18.059 20:15:13 -- common/autotest_common.sh@1564 -- # cat /sys/bus/pci/devices/0000:00:10.0/device 00:16:18.059 20:15:13 -- common/autotest_common.sh@1564 -- # device=0x0010 00:16:18.059 20:15:13 -- common/autotest_common.sh@1565 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:16:18.059 20:15:13 -- common/autotest_common.sh@1563 -- # for bdf in "${_bdfs[@]}" 00:16:18.059 20:15:13 -- common/autotest_common.sh@1564 -- # cat /sys/bus/pci/devices/0000:00:11.0/device 00:16:18.059 20:15:13 -- common/autotest_common.sh@1564 -- # device=0x0010 00:16:18.059 20:15:13 -- common/autotest_common.sh@1565 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:16:18.059 20:15:13 -- common/autotest_common.sh@1570 -- # (( 0 > 0 )) 00:16:18.059 20:15:13 -- common/autotest_common.sh@1570 -- # return 0 00:16:18.059 20:15:13 -- common/autotest_common.sh@1577 -- # [[ -z '' ]] 00:16:18.059 20:15:13 -- common/autotest_common.sh@1578 -- # return 0 00:16:18.059 20:15:13 -- spdk/autotest.sh@137 -- # '[' 0 -eq 1 ']' 00:16:18.059 20:15:13 -- spdk/autotest.sh@141 -- # '[' 1 -eq 1 ']' 00:16:18.059 20:15:13 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:16:18.059 20:15:13 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:16:18.059 20:15:13 -- spdk/autotest.sh@149 -- # timing_enter lib 00:16:18.059 20:15:13 -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:18.059 20:15:13 -- common/autotest_common.sh@10 -- # set +x 00:16:18.059 20:15:13 -- spdk/autotest.sh@151 -- # [[ 0 -eq 1 ]] 00:16:18.059 20:15:13 -- spdk/autotest.sh@155 -- # run_test env /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:16:18.059 20:15:13 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:18.059 20:15:13 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:18.059 20:15:13 -- common/autotest_common.sh@10 -- # set +x 00:16:18.059 ************************************ 00:16:18.059 START TEST env 00:16:18.059 ************************************ 00:16:18.059 20:15:13 env -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:16:18.317 * Looking for test storage... 00:16:18.317 * Found test storage at /home/vagrant/spdk_repo/spdk/test/env 00:16:18.317 20:15:13 env -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:16:18.317 20:15:13 env -- common/autotest_common.sh@1681 -- # lcov --version 00:16:18.317 20:15:13 env -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:16:18.317 20:15:13 env -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:16:18.317 20:15:13 env -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:18.317 20:15:13 env -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:18.317 20:15:13 env -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:18.317 20:15:13 env -- scripts/common.sh@336 -- # IFS=.-: 00:16:18.318 20:15:13 env -- scripts/common.sh@336 -- # read -ra ver1 00:16:18.318 20:15:13 env -- scripts/common.sh@337 -- # IFS=.-: 00:16:18.318 20:15:13 env -- scripts/common.sh@337 -- # read -ra ver2 00:16:18.318 20:15:13 env -- scripts/common.sh@338 -- # local 'op=<' 00:16:18.318 20:15:13 env -- scripts/common.sh@340 -- # ver1_l=2 00:16:18.318 20:15:13 env -- scripts/common.sh@341 -- # ver2_l=1 00:16:18.318 20:15:13 env -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:18.318 20:15:13 env -- scripts/common.sh@344 -- # case "$op" in 00:16:18.318 20:15:13 env -- scripts/common.sh@345 -- # : 1 00:16:18.318 20:15:13 env -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:18.318 20:15:13 env -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:18.318 20:15:13 env -- scripts/common.sh@365 -- # decimal 1 00:16:18.318 20:15:13 env -- scripts/common.sh@353 -- # local d=1 00:16:18.318 20:15:13 env -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:18.318 20:15:13 env -- scripts/common.sh@355 -- # echo 1 00:16:18.318 20:15:13 env -- scripts/common.sh@365 -- # ver1[v]=1 00:16:18.318 20:15:13 env -- scripts/common.sh@366 -- # decimal 2 00:16:18.318 20:15:13 env -- scripts/common.sh@353 -- # local d=2 00:16:18.318 20:15:13 env -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:18.318 20:15:13 env -- scripts/common.sh@355 -- # echo 2 00:16:18.318 20:15:13 env -- scripts/common.sh@366 -- # ver2[v]=2 00:16:18.318 20:15:13 env -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:18.318 20:15:13 env -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:18.318 20:15:13 env -- scripts/common.sh@368 -- # return 0 00:16:18.318 20:15:13 env -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:18.318 20:15:13 env -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:16:18.318 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:18.318 --rc genhtml_branch_coverage=1 00:16:18.318 --rc genhtml_function_coverage=1 00:16:18.318 --rc genhtml_legend=1 00:16:18.318 --rc geninfo_all_blocks=1 00:16:18.318 --rc geninfo_unexecuted_blocks=1 00:16:18.318 00:16:18.318 ' 00:16:18.318 20:15:13 env -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:16:18.318 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:18.318 --rc genhtml_branch_coverage=1 00:16:18.318 --rc genhtml_function_coverage=1 00:16:18.318 --rc genhtml_legend=1 00:16:18.318 --rc geninfo_all_blocks=1 00:16:18.318 --rc geninfo_unexecuted_blocks=1 00:16:18.318 00:16:18.318 ' 00:16:18.318 20:15:13 env -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:16:18.318 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:18.318 --rc genhtml_branch_coverage=1 00:16:18.318 --rc genhtml_function_coverage=1 00:16:18.318 --rc genhtml_legend=1 00:16:18.318 --rc geninfo_all_blocks=1 00:16:18.318 --rc geninfo_unexecuted_blocks=1 00:16:18.318 00:16:18.318 ' 00:16:18.318 20:15:13 env -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:16:18.318 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:18.318 --rc genhtml_branch_coverage=1 00:16:18.318 --rc genhtml_function_coverage=1 00:16:18.318 --rc genhtml_legend=1 00:16:18.318 --rc geninfo_all_blocks=1 00:16:18.318 --rc geninfo_unexecuted_blocks=1 00:16:18.318 00:16:18.318 ' 00:16:18.318 20:15:13 env -- env/env.sh@10 -- # run_test env_memory /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:16:18.318 20:15:13 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:18.318 20:15:13 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:18.318 20:15:13 env -- common/autotest_common.sh@10 -- # set +x 00:16:18.318 ************************************ 00:16:18.318 START TEST env_memory 00:16:18.318 ************************************ 00:16:18.318 20:15:13 env.env_memory -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:16:18.318 00:16:18.318 00:16:18.318 CUnit - A unit testing framework for C - Version 2.1-3 00:16:18.318 http://cunit.sourceforge.net/ 00:16:18.318 00:16:18.318 00:16:18.318 Suite: memory 00:16:18.318 Test: alloc and free memory map ...[2024-10-01 20:15:13.550433] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:16:18.577 passed 00:16:18.577 Test: mem map translation ...[2024-10-01 20:15:13.611180] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:16:18.577 [2024-10-01 20:15:13.611306] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:16:18.577 [2024-10-01 20:15:13.611410] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 589:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:16:18.577 [2024-10-01 20:15:13.611445] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 605:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:16:18.577 passed 00:16:18.577 Test: mem map registration ...[2024-10-01 20:15:13.710676] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=200000 len=1234 00:16:18.577 [2024-10-01 20:15:13.710832] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=4d2 len=2097152 00:16:18.577 passed 00:16:18.835 Test: mem map adjacent registrations ...passed 00:16:18.835 00:16:18.835 Run Summary: Type Total Ran Passed Failed Inactive 00:16:18.835 suites 1 1 n/a 0 0 00:16:18.835 tests 4 4 4 0 0 00:16:18.835 asserts 152 152 152 0 n/a 00:16:18.835 00:16:18.835 Elapsed time = 0.346 seconds 00:16:18.835 00:16:18.835 real 0m0.390s 00:16:18.835 user 0m0.360s 00:16:18.835 sys 0m0.022s 00:16:18.835 20:15:13 env.env_memory -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:18.835 ************************************ 00:16:18.835 END TEST env_memory 00:16:18.835 ************************************ 00:16:18.835 20:15:13 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:16:18.835 20:15:13 env -- env/env.sh@11 -- # run_test env_vtophys /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:16:18.835 20:15:13 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:18.835 20:15:13 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:18.835 20:15:13 env -- common/autotest_common.sh@10 -- # set +x 00:16:18.835 ************************************ 00:16:18.835 START TEST env_vtophys 00:16:18.835 ************************************ 00:16:18.835 20:15:13 env.env_vtophys -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:16:18.835 EAL: lib.eal log level changed from notice to debug 00:16:18.835 EAL: Detected lcore 0 as core 0 on socket 0 00:16:18.835 EAL: Detected lcore 1 as core 0 on socket 0 00:16:18.835 EAL: Detected lcore 2 as core 0 on socket 0 00:16:18.835 EAL: Detected lcore 3 as core 0 on socket 0 00:16:18.835 EAL: Detected lcore 4 as core 0 on socket 0 00:16:18.835 EAL: Detected lcore 5 as core 0 on socket 0 00:16:18.835 EAL: Detected lcore 6 as core 0 on socket 0 00:16:18.835 EAL: Detected lcore 7 as core 0 on socket 0 00:16:18.835 EAL: Detected lcore 8 as core 0 on socket 0 00:16:18.835 EAL: Detected lcore 9 as core 0 on socket 0 00:16:18.835 EAL: Maximum logical cores by configuration: 128 00:16:18.835 EAL: Detected CPU lcores: 10 00:16:18.835 EAL: Detected NUMA nodes: 1 00:16:18.835 EAL: Checking presence of .so 'librte_eal.so.24.1' 00:16:18.835 EAL: Detected shared linkage of DPDK 00:16:18.835 EAL: No shared files mode enabled, IPC will be disabled 00:16:18.835 EAL: Selected IOVA mode 'PA' 00:16:18.835 EAL: Probing VFIO support... 00:16:18.835 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:16:18.835 EAL: VFIO modules not loaded, skipping VFIO support... 00:16:18.835 EAL: Ask a virtual area of 0x2e000 bytes 00:16:18.835 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:16:18.835 EAL: Setting up physically contiguous memory... 00:16:18.835 EAL: Setting maximum number of open files to 524288 00:16:18.835 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:16:18.835 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:16:18.835 EAL: Ask a virtual area of 0x61000 bytes 00:16:18.835 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:16:18.835 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:16:18.835 EAL: Ask a virtual area of 0x400000000 bytes 00:16:18.835 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:16:18.835 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:16:18.835 EAL: Ask a virtual area of 0x61000 bytes 00:16:18.835 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:16:18.835 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:16:18.835 EAL: Ask a virtual area of 0x400000000 bytes 00:16:18.835 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:16:18.835 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:16:18.835 EAL: Ask a virtual area of 0x61000 bytes 00:16:18.835 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:16:18.835 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:16:18.835 EAL: Ask a virtual area of 0x400000000 bytes 00:16:18.835 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:16:18.835 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:16:18.835 EAL: Ask a virtual area of 0x61000 bytes 00:16:18.835 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:16:18.835 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:16:18.835 EAL: Ask a virtual area of 0x400000000 bytes 00:16:18.835 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:16:18.835 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:16:18.835 EAL: Hugepages will be freed exactly as allocated. 00:16:18.835 EAL: No shared files mode enabled, IPC is disabled 00:16:18.835 EAL: No shared files mode enabled, IPC is disabled 00:16:19.170 EAL: TSC frequency is ~2200000 KHz 00:16:19.170 EAL: Main lcore 0 is ready (tid=7f23bd611a40;cpuset=[0]) 00:16:19.170 EAL: Trying to obtain current memory policy. 00:16:19.170 EAL: Setting policy MPOL_PREFERRED for socket 0 00:16:19.170 EAL: Restoring previous memory policy: 0 00:16:19.170 EAL: request: mp_malloc_sync 00:16:19.170 EAL: No shared files mode enabled, IPC is disabled 00:16:19.170 EAL: Heap on socket 0 was expanded by 2MB 00:16:19.170 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:16:19.170 EAL: No PCI address specified using 'addr=' in: bus=pci 00:16:19.170 EAL: Mem event callback 'spdk:(nil)' registered 00:16:19.170 EAL: Module /sys/module/vfio_pci not found! error 2 (No such file or directory) 00:16:19.170 00:16:19.170 00:16:19.170 CUnit - A unit testing framework for C - Version 2.1-3 00:16:19.170 http://cunit.sourceforge.net/ 00:16:19.170 00:16:19.170 00:16:19.170 Suite: components_suite 00:16:19.428 Test: vtophys_malloc_test ...passed 00:16:19.428 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:16:19.428 EAL: Setting policy MPOL_PREFERRED for socket 0 00:16:19.428 EAL: Restoring previous memory policy: 4 00:16:19.428 EAL: Calling mem event callback 'spdk:(nil)' 00:16:19.428 EAL: request: mp_malloc_sync 00:16:19.428 EAL: No shared files mode enabled, IPC is disabled 00:16:19.428 EAL: Heap on socket 0 was expanded by 4MB 00:16:19.428 EAL: Calling mem event callback 'spdk:(nil)' 00:16:19.428 EAL: request: mp_malloc_sync 00:16:19.428 EAL: No shared files mode enabled, IPC is disabled 00:16:19.428 EAL: Heap on socket 0 was shrunk by 4MB 00:16:19.428 EAL: Trying to obtain current memory policy. 00:16:19.428 EAL: Setting policy MPOL_PREFERRED for socket 0 00:16:19.428 EAL: Restoring previous memory policy: 4 00:16:19.428 EAL: Calling mem event callback 'spdk:(nil)' 00:16:19.428 EAL: request: mp_malloc_sync 00:16:19.428 EAL: No shared files mode enabled, IPC is disabled 00:16:19.428 EAL: Heap on socket 0 was expanded by 6MB 00:16:19.428 EAL: Calling mem event callback 'spdk:(nil)' 00:16:19.428 EAL: request: mp_malloc_sync 00:16:19.428 EAL: No shared files mode enabled, IPC is disabled 00:16:19.428 EAL: Heap on socket 0 was shrunk by 6MB 00:16:19.428 EAL: Trying to obtain current memory policy. 00:16:19.428 EAL: Setting policy MPOL_PREFERRED for socket 0 00:16:19.428 EAL: Restoring previous memory policy: 4 00:16:19.428 EAL: Calling mem event callback 'spdk:(nil)' 00:16:19.428 EAL: request: mp_malloc_sync 00:16:19.428 EAL: No shared files mode enabled, IPC is disabled 00:16:19.428 EAL: Heap on socket 0 was expanded by 10MB 00:16:19.686 EAL: Calling mem event callback 'spdk:(nil)' 00:16:19.686 EAL: request: mp_malloc_sync 00:16:19.686 EAL: No shared files mode enabled, IPC is disabled 00:16:19.686 EAL: Heap on socket 0 was shrunk by 10MB 00:16:19.686 EAL: Trying to obtain current memory policy. 00:16:19.686 EAL: Setting policy MPOL_PREFERRED for socket 0 00:16:19.686 EAL: Restoring previous memory policy: 4 00:16:19.686 EAL: Calling mem event callback 'spdk:(nil)' 00:16:19.686 EAL: request: mp_malloc_sync 00:16:19.686 EAL: No shared files mode enabled, IPC is disabled 00:16:19.686 EAL: Heap on socket 0 was expanded by 18MB 00:16:19.686 EAL: Calling mem event callback 'spdk:(nil)' 00:16:19.686 EAL: request: mp_malloc_sync 00:16:19.686 EAL: No shared files mode enabled, IPC is disabled 00:16:19.686 EAL: Heap on socket 0 was shrunk by 18MB 00:16:19.686 EAL: Trying to obtain current memory policy. 00:16:19.686 EAL: Setting policy MPOL_PREFERRED for socket 0 00:16:19.686 EAL: Restoring previous memory policy: 4 00:16:19.686 EAL: Calling mem event callback 'spdk:(nil)' 00:16:19.686 EAL: request: mp_malloc_sync 00:16:19.686 EAL: No shared files mode enabled, IPC is disabled 00:16:19.686 EAL: Heap on socket 0 was expanded by 34MB 00:16:19.686 EAL: Calling mem event callback 'spdk:(nil)' 00:16:19.686 EAL: request: mp_malloc_sync 00:16:19.686 EAL: No shared files mode enabled, IPC is disabled 00:16:19.687 EAL: Heap on socket 0 was shrunk by 34MB 00:16:19.687 EAL: Trying to obtain current memory policy. 00:16:19.687 EAL: Setting policy MPOL_PREFERRED for socket 0 00:16:19.687 EAL: Restoring previous memory policy: 4 00:16:19.687 EAL: Calling mem event callback 'spdk:(nil)' 00:16:19.687 EAL: request: mp_malloc_sync 00:16:19.687 EAL: No shared files mode enabled, IPC is disabled 00:16:19.687 EAL: Heap on socket 0 was expanded by 66MB 00:16:19.946 EAL: Calling mem event callback 'spdk:(nil)' 00:16:19.946 EAL: request: mp_malloc_sync 00:16:19.946 EAL: No shared files mode enabled, IPC is disabled 00:16:19.946 EAL: Heap on socket 0 was shrunk by 66MB 00:16:19.946 EAL: Trying to obtain current memory policy. 00:16:19.946 EAL: Setting policy MPOL_PREFERRED for socket 0 00:16:19.946 EAL: Restoring previous memory policy: 4 00:16:19.946 EAL: Calling mem event callback 'spdk:(nil)' 00:16:19.946 EAL: request: mp_malloc_sync 00:16:19.946 EAL: No shared files mode enabled, IPC is disabled 00:16:19.946 EAL: Heap on socket 0 was expanded by 130MB 00:16:20.204 EAL: Calling mem event callback 'spdk:(nil)' 00:16:20.204 EAL: request: mp_malloc_sync 00:16:20.204 EAL: No shared files mode enabled, IPC is disabled 00:16:20.204 EAL: Heap on socket 0 was shrunk by 130MB 00:16:20.463 EAL: Trying to obtain current memory policy. 00:16:20.463 EAL: Setting policy MPOL_PREFERRED for socket 0 00:16:20.463 EAL: Restoring previous memory policy: 4 00:16:20.463 EAL: Calling mem event callback 'spdk:(nil)' 00:16:20.463 EAL: request: mp_malloc_sync 00:16:20.463 EAL: No shared files mode enabled, IPC is disabled 00:16:20.463 EAL: Heap on socket 0 was expanded by 258MB 00:16:21.032 EAL: Calling mem event callback 'spdk:(nil)' 00:16:21.032 EAL: request: mp_malloc_sync 00:16:21.032 EAL: No shared files mode enabled, IPC is disabled 00:16:21.032 EAL: Heap on socket 0 was shrunk by 258MB 00:16:21.290 EAL: Trying to obtain current memory policy. 00:16:21.290 EAL: Setting policy MPOL_PREFERRED for socket 0 00:16:21.549 EAL: Restoring previous memory policy: 4 00:16:21.549 EAL: Calling mem event callback 'spdk:(nil)' 00:16:21.549 EAL: request: mp_malloc_sync 00:16:21.549 EAL: No shared files mode enabled, IPC is disabled 00:16:21.549 EAL: Heap on socket 0 was expanded by 514MB 00:16:22.486 EAL: Calling mem event callback 'spdk:(nil)' 00:16:22.486 EAL: request: mp_malloc_sync 00:16:22.486 EAL: No shared files mode enabled, IPC is disabled 00:16:22.486 EAL: Heap on socket 0 was shrunk by 514MB 00:16:23.064 EAL: Trying to obtain current memory policy. 00:16:23.064 EAL: Setting policy MPOL_PREFERRED for socket 0 00:16:23.323 EAL: Restoring previous memory policy: 4 00:16:23.323 EAL: Calling mem event callback 'spdk:(nil)' 00:16:23.323 EAL: request: mp_malloc_sync 00:16:23.323 EAL: No shared files mode enabled, IPC is disabled 00:16:23.323 EAL: Heap on socket 0 was expanded by 1026MB 00:16:25.224 EAL: Calling mem event callback 'spdk:(nil)' 00:16:25.224 EAL: request: mp_malloc_sync 00:16:25.224 EAL: No shared files mode enabled, IPC is disabled 00:16:25.224 EAL: Heap on socket 0 was shrunk by 1026MB 00:16:26.604 passed 00:16:26.604 00:16:26.604 Run Summary: Type Total Ran Passed Failed Inactive 00:16:26.604 suites 1 1 n/a 0 0 00:16:26.604 tests 2 2 2 0 0 00:16:26.604 asserts 5663 5663 5663 0 n/a 00:16:26.604 00:16:26.604 Elapsed time = 7.549 seconds 00:16:26.604 EAL: Calling mem event callback 'spdk:(nil)' 00:16:26.604 EAL: request: mp_malloc_sync 00:16:26.604 EAL: No shared files mode enabled, IPC is disabled 00:16:26.604 EAL: Heap on socket 0 was shrunk by 2MB 00:16:26.604 EAL: No shared files mode enabled, IPC is disabled 00:16:26.604 EAL: No shared files mode enabled, IPC is disabled 00:16:26.604 EAL: No shared files mode enabled, IPC is disabled 00:16:26.604 00:16:26.604 real 0m7.887s 00:16:26.604 user 0m6.639s 00:16:26.604 sys 0m1.079s 00:16:26.604 20:15:21 env.env_vtophys -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:26.604 20:15:21 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:16:26.604 ************************************ 00:16:26.604 END TEST env_vtophys 00:16:26.604 ************************************ 00:16:26.604 20:15:21 env -- env/env.sh@12 -- # run_test env_pci /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:16:26.604 20:15:21 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:26.604 20:15:21 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:26.604 20:15:21 env -- common/autotest_common.sh@10 -- # set +x 00:16:26.863 ************************************ 00:16:26.863 START TEST env_pci 00:16:26.863 ************************************ 00:16:26.863 20:15:21 env.env_pci -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:16:26.863 00:16:26.863 00:16:26.863 CUnit - A unit testing framework for C - Version 2.1-3 00:16:26.863 http://cunit.sourceforge.net/ 00:16:26.863 00:16:26.863 00:16:26.863 Suite: pci 00:16:26.863 Test: pci_hook ...[2024-10-01 20:15:21.898063] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/pci.c:1049:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 56704 has claimed it 00:16:26.863 EAL: Cannot find device (10000:00:01.0) 00:16:26.863 EAL: Failed to attach device on primary process 00:16:26.863 passed 00:16:26.863 00:16:26.863 Run Summary: Type Total Ran Passed Failed Inactive 00:16:26.863 suites 1 1 n/a 0 0 00:16:26.863 tests 1 1 1 0 0 00:16:26.863 asserts 25 25 25 0 n/a 00:16:26.863 00:16:26.863 Elapsed time = 0.008 seconds 00:16:26.863 00:16:26.863 real 0m0.079s 00:16:26.863 user 0m0.040s 00:16:26.863 sys 0m0.038s 00:16:26.863 20:15:21 env.env_pci -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:26.863 ************************************ 00:16:26.863 END TEST env_pci 00:16:26.863 ************************************ 00:16:26.863 20:15:21 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:16:26.863 20:15:21 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:16:26.863 20:15:21 env -- env/env.sh@15 -- # uname 00:16:26.863 20:15:21 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:16:26.863 20:15:21 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:16:26.863 20:15:21 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:16:26.863 20:15:21 env -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:16:26.863 20:15:21 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:26.863 20:15:21 env -- common/autotest_common.sh@10 -- # set +x 00:16:26.863 ************************************ 00:16:26.863 START TEST env_dpdk_post_init 00:16:26.863 ************************************ 00:16:26.863 20:15:22 env.env_dpdk_post_init -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:16:26.863 EAL: Detected CPU lcores: 10 00:16:26.863 EAL: Detected NUMA nodes: 1 00:16:26.863 EAL: Detected shared linkage of DPDK 00:16:26.863 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:16:26.863 EAL: Selected IOVA mode 'PA' 00:16:27.122 TELEMETRY: No legacy callbacks, legacy socket not created 00:16:27.122 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:10.0 (socket -1) 00:16:27.122 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:11.0 (socket -1) 00:16:27.122 Starting DPDK initialization... 00:16:27.122 Starting SPDK post initialization... 00:16:27.122 SPDK NVMe probe 00:16:27.122 Attaching to 0000:00:10.0 00:16:27.122 Attaching to 0000:00:11.0 00:16:27.122 Attached to 0000:00:10.0 00:16:27.122 Attached to 0000:00:11.0 00:16:27.122 Cleaning up... 00:16:27.122 00:16:27.122 real 0m0.289s 00:16:27.122 user 0m0.086s 00:16:27.122 sys 0m0.103s 00:16:27.122 20:15:22 env.env_dpdk_post_init -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:27.122 ************************************ 00:16:27.122 END TEST env_dpdk_post_init 00:16:27.122 ************************************ 00:16:27.122 20:15:22 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:16:27.122 20:15:22 env -- env/env.sh@26 -- # uname 00:16:27.122 20:15:22 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:16:27.122 20:15:22 env -- env/env.sh@29 -- # run_test env_mem_callbacks /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:16:27.122 20:15:22 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:27.122 20:15:22 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:27.122 20:15:22 env -- common/autotest_common.sh@10 -- # set +x 00:16:27.122 ************************************ 00:16:27.122 START TEST env_mem_callbacks 00:16:27.122 ************************************ 00:16:27.122 20:15:22 env.env_mem_callbacks -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:16:27.380 EAL: Detected CPU lcores: 10 00:16:27.380 EAL: Detected NUMA nodes: 1 00:16:27.380 EAL: Detected shared linkage of DPDK 00:16:27.380 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:16:27.380 EAL: Selected IOVA mode 'PA' 00:16:27.380 00:16:27.380 00:16:27.380 CUnit - A unit testing framework for C - Version 2.1-3 00:16:27.380 http://cunit.sourceforge.net/ 00:16:27.380 00:16:27.380 00:16:27.380 Suite: memory 00:16:27.380 Test: test ... 00:16:27.380 register 0x200000200000 2097152 00:16:27.380 malloc 3145728 00:16:27.380 TELEMETRY: No legacy callbacks, legacy socket not created 00:16:27.380 register 0x200000400000 4194304 00:16:27.380 buf 0x2000004fffc0 len 3145728 PASSED 00:16:27.380 malloc 64 00:16:27.380 buf 0x2000004ffec0 len 64 PASSED 00:16:27.380 malloc 4194304 00:16:27.380 register 0x200000800000 6291456 00:16:27.380 buf 0x2000009fffc0 len 4194304 PASSED 00:16:27.380 free 0x2000004fffc0 3145728 00:16:27.380 free 0x2000004ffec0 64 00:16:27.380 unregister 0x200000400000 4194304 PASSED 00:16:27.380 free 0x2000009fffc0 4194304 00:16:27.380 unregister 0x200000800000 6291456 PASSED 00:16:27.380 malloc 8388608 00:16:27.380 register 0x200000400000 10485760 00:16:27.380 buf 0x2000005fffc0 len 8388608 PASSED 00:16:27.380 free 0x2000005fffc0 8388608 00:16:27.380 unregister 0x200000400000 10485760 PASSED 00:16:27.380 passed 00:16:27.380 00:16:27.380 Run Summary: Type Total Ran Passed Failed Inactive 00:16:27.380 suites 1 1 n/a 0 0 00:16:27.380 tests 1 1 1 0 0 00:16:27.380 asserts 15 15 15 0 n/a 00:16:27.380 00:16:27.380 Elapsed time = 0.072 seconds 00:16:27.639 00:16:27.639 real 0m0.287s 00:16:27.639 user 0m0.120s 00:16:27.639 sys 0m0.064s 00:16:27.639 20:15:22 env.env_mem_callbacks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:27.639 20:15:22 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:16:27.639 ************************************ 00:16:27.639 END TEST env_mem_callbacks 00:16:27.639 ************************************ 00:16:27.639 00:16:27.639 real 0m9.419s 00:16:27.639 user 0m7.471s 00:16:27.639 sys 0m1.558s 00:16:27.639 20:15:22 env -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:27.639 20:15:22 env -- common/autotest_common.sh@10 -- # set +x 00:16:27.639 ************************************ 00:16:27.639 END TEST env 00:16:27.639 ************************************ 00:16:27.639 20:15:22 -- spdk/autotest.sh@156 -- # run_test rpc /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:16:27.639 20:15:22 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:27.639 20:15:22 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:27.639 20:15:22 -- common/autotest_common.sh@10 -- # set +x 00:16:27.639 ************************************ 00:16:27.639 START TEST rpc 00:16:27.639 ************************************ 00:16:27.639 20:15:22 rpc -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:16:27.639 * Looking for test storage... 00:16:27.639 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:16:27.639 20:15:22 rpc -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:16:27.639 20:15:22 rpc -- common/autotest_common.sh@1681 -- # lcov --version 00:16:27.639 20:15:22 rpc -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:16:27.897 20:15:22 rpc -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:16:27.897 20:15:22 rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:27.897 20:15:22 rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:27.897 20:15:22 rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:27.897 20:15:22 rpc -- scripts/common.sh@336 -- # IFS=.-: 00:16:27.897 20:15:22 rpc -- scripts/common.sh@336 -- # read -ra ver1 00:16:27.897 20:15:22 rpc -- scripts/common.sh@337 -- # IFS=.-: 00:16:27.897 20:15:22 rpc -- scripts/common.sh@337 -- # read -ra ver2 00:16:27.897 20:15:22 rpc -- scripts/common.sh@338 -- # local 'op=<' 00:16:27.897 20:15:22 rpc -- scripts/common.sh@340 -- # ver1_l=2 00:16:27.897 20:15:22 rpc -- scripts/common.sh@341 -- # ver2_l=1 00:16:27.897 20:15:22 rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:27.897 20:15:22 rpc -- scripts/common.sh@344 -- # case "$op" in 00:16:27.897 20:15:22 rpc -- scripts/common.sh@345 -- # : 1 00:16:27.897 20:15:22 rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:27.897 20:15:22 rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:27.897 20:15:22 rpc -- scripts/common.sh@365 -- # decimal 1 00:16:27.897 20:15:22 rpc -- scripts/common.sh@353 -- # local d=1 00:16:27.897 20:15:22 rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:27.897 20:15:22 rpc -- scripts/common.sh@355 -- # echo 1 00:16:27.897 20:15:22 rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:16:27.897 20:15:22 rpc -- scripts/common.sh@366 -- # decimal 2 00:16:27.897 20:15:22 rpc -- scripts/common.sh@353 -- # local d=2 00:16:27.897 20:15:22 rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:27.897 20:15:22 rpc -- scripts/common.sh@355 -- # echo 2 00:16:27.897 20:15:22 rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:16:27.897 20:15:22 rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:27.897 20:15:22 rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:27.897 20:15:22 rpc -- scripts/common.sh@368 -- # return 0 00:16:27.897 20:15:22 rpc -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:27.897 20:15:22 rpc -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:16:27.897 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:27.897 --rc genhtml_branch_coverage=1 00:16:27.897 --rc genhtml_function_coverage=1 00:16:27.897 --rc genhtml_legend=1 00:16:27.897 --rc geninfo_all_blocks=1 00:16:27.897 --rc geninfo_unexecuted_blocks=1 00:16:27.898 00:16:27.898 ' 00:16:27.898 20:15:22 rpc -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:16:27.898 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:27.898 --rc genhtml_branch_coverage=1 00:16:27.898 --rc genhtml_function_coverage=1 00:16:27.898 --rc genhtml_legend=1 00:16:27.898 --rc geninfo_all_blocks=1 00:16:27.898 --rc geninfo_unexecuted_blocks=1 00:16:27.898 00:16:27.898 ' 00:16:27.898 20:15:22 rpc -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:16:27.898 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:27.898 --rc genhtml_branch_coverage=1 00:16:27.898 --rc genhtml_function_coverage=1 00:16:27.898 --rc genhtml_legend=1 00:16:27.898 --rc geninfo_all_blocks=1 00:16:27.898 --rc geninfo_unexecuted_blocks=1 00:16:27.898 00:16:27.898 ' 00:16:27.898 20:15:22 rpc -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:16:27.898 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:27.898 --rc genhtml_branch_coverage=1 00:16:27.898 --rc genhtml_function_coverage=1 00:16:27.898 --rc genhtml_legend=1 00:16:27.898 --rc geninfo_all_blocks=1 00:16:27.898 --rc geninfo_unexecuted_blocks=1 00:16:27.898 00:16:27.898 ' 00:16:27.898 20:15:22 rpc -- rpc/rpc.sh@65 -- # spdk_pid=56831 00:16:27.898 20:15:22 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:16:27.898 20:15:22 rpc -- rpc/rpc.sh@64 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -e bdev 00:16:27.898 20:15:22 rpc -- rpc/rpc.sh@67 -- # waitforlisten 56831 00:16:27.898 20:15:22 rpc -- common/autotest_common.sh@831 -- # '[' -z 56831 ']' 00:16:27.898 20:15:22 rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:27.898 20:15:22 rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:27.898 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:27.898 20:15:22 rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:27.898 20:15:22 rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:27.898 20:15:22 rpc -- common/autotest_common.sh@10 -- # set +x 00:16:27.898 [2024-10-01 20:15:23.114151] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:16:27.898 [2024-10-01 20:15:23.114356] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid56831 ] 00:16:28.156 [2024-10-01 20:15:23.297319] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:28.415 [2024-10-01 20:15:23.584411] app.c: 610:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:16:28.415 [2024-10-01 20:15:23.584514] app.c: 611:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 56831' to capture a snapshot of events at runtime. 00:16:28.415 [2024-10-01 20:15:23.584534] app.c: 616:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:16:28.415 [2024-10-01 20:15:23.584551] app.c: 617:app_setup_trace: *NOTICE*: SPDK application currently running. 00:16:28.415 [2024-10-01 20:15:23.584564] app.c: 618:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid56831 for offline analysis/debug. 00:16:28.415 [2024-10-01 20:15:23.584619] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:29.790 20:15:24 rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:29.790 20:15:24 rpc -- common/autotest_common.sh@864 -- # return 0 00:16:29.790 20:15:24 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:16:29.790 20:15:24 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:16:29.790 20:15:24 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:16:29.790 20:15:24 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:16:29.790 20:15:24 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:29.790 20:15:24 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:29.791 20:15:24 rpc -- common/autotest_common.sh@10 -- # set +x 00:16:29.791 ************************************ 00:16:29.791 START TEST rpc_integrity 00:16:29.791 ************************************ 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@1125 -- # rpc_integrity 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:16:29.791 { 00:16:29.791 "name": "Malloc0", 00:16:29.791 "aliases": [ 00:16:29.791 "c49c735c-d929-4a2c-872b-a1b3494b4aac" 00:16:29.791 ], 00:16:29.791 "product_name": "Malloc disk", 00:16:29.791 "block_size": 512, 00:16:29.791 "num_blocks": 16384, 00:16:29.791 "uuid": "c49c735c-d929-4a2c-872b-a1b3494b4aac", 00:16:29.791 "assigned_rate_limits": { 00:16:29.791 "rw_ios_per_sec": 0, 00:16:29.791 "rw_mbytes_per_sec": 0, 00:16:29.791 "r_mbytes_per_sec": 0, 00:16:29.791 "w_mbytes_per_sec": 0 00:16:29.791 }, 00:16:29.791 "claimed": false, 00:16:29.791 "zoned": false, 00:16:29.791 "supported_io_types": { 00:16:29.791 "read": true, 00:16:29.791 "write": true, 00:16:29.791 "unmap": true, 00:16:29.791 "flush": true, 00:16:29.791 "reset": true, 00:16:29.791 "nvme_admin": false, 00:16:29.791 "nvme_io": false, 00:16:29.791 "nvme_io_md": false, 00:16:29.791 "write_zeroes": true, 00:16:29.791 "zcopy": true, 00:16:29.791 "get_zone_info": false, 00:16:29.791 "zone_management": false, 00:16:29.791 "zone_append": false, 00:16:29.791 "compare": false, 00:16:29.791 "compare_and_write": false, 00:16:29.791 "abort": true, 00:16:29.791 "seek_hole": false, 00:16:29.791 "seek_data": false, 00:16:29.791 "copy": true, 00:16:29.791 "nvme_iov_md": false 00:16:29.791 }, 00:16:29.791 "memory_domains": [ 00:16:29.791 { 00:16:29.791 "dma_device_id": "system", 00:16:29.791 "dma_device_type": 1 00:16:29.791 }, 00:16:29.791 { 00:16:29.791 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:29.791 "dma_device_type": 2 00:16:29.791 } 00:16:29.791 ], 00:16:29.791 "driver_specific": {} 00:16:29.791 } 00:16:29.791 ]' 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:29.791 [2024-10-01 20:15:24.894375] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:16:29.791 [2024-10-01 20:15:24.894470] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:29.791 [2024-10-01 20:15:24.894511] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:16:29.791 [2024-10-01 20:15:24.894532] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:29.791 [2024-10-01 20:15:24.897911] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:29.791 [2024-10-01 20:15:24.897967] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:16:29.791 Passthru0 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:16:29.791 { 00:16:29.791 "name": "Malloc0", 00:16:29.791 "aliases": [ 00:16:29.791 "c49c735c-d929-4a2c-872b-a1b3494b4aac" 00:16:29.791 ], 00:16:29.791 "product_name": "Malloc disk", 00:16:29.791 "block_size": 512, 00:16:29.791 "num_blocks": 16384, 00:16:29.791 "uuid": "c49c735c-d929-4a2c-872b-a1b3494b4aac", 00:16:29.791 "assigned_rate_limits": { 00:16:29.791 "rw_ios_per_sec": 0, 00:16:29.791 "rw_mbytes_per_sec": 0, 00:16:29.791 "r_mbytes_per_sec": 0, 00:16:29.791 "w_mbytes_per_sec": 0 00:16:29.791 }, 00:16:29.791 "claimed": true, 00:16:29.791 "claim_type": "exclusive_write", 00:16:29.791 "zoned": false, 00:16:29.791 "supported_io_types": { 00:16:29.791 "read": true, 00:16:29.791 "write": true, 00:16:29.791 "unmap": true, 00:16:29.791 "flush": true, 00:16:29.791 "reset": true, 00:16:29.791 "nvme_admin": false, 00:16:29.791 "nvme_io": false, 00:16:29.791 "nvme_io_md": false, 00:16:29.791 "write_zeroes": true, 00:16:29.791 "zcopy": true, 00:16:29.791 "get_zone_info": false, 00:16:29.791 "zone_management": false, 00:16:29.791 "zone_append": false, 00:16:29.791 "compare": false, 00:16:29.791 "compare_and_write": false, 00:16:29.791 "abort": true, 00:16:29.791 "seek_hole": false, 00:16:29.791 "seek_data": false, 00:16:29.791 "copy": true, 00:16:29.791 "nvme_iov_md": false 00:16:29.791 }, 00:16:29.791 "memory_domains": [ 00:16:29.791 { 00:16:29.791 "dma_device_id": "system", 00:16:29.791 "dma_device_type": 1 00:16:29.791 }, 00:16:29.791 { 00:16:29.791 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:29.791 "dma_device_type": 2 00:16:29.791 } 00:16:29.791 ], 00:16:29.791 "driver_specific": {} 00:16:29.791 }, 00:16:29.791 { 00:16:29.791 "name": "Passthru0", 00:16:29.791 "aliases": [ 00:16:29.791 "74a71256-5bc7-5f4b-9654-87c03f9f30ce" 00:16:29.791 ], 00:16:29.791 "product_name": "passthru", 00:16:29.791 "block_size": 512, 00:16:29.791 "num_blocks": 16384, 00:16:29.791 "uuid": "74a71256-5bc7-5f4b-9654-87c03f9f30ce", 00:16:29.791 "assigned_rate_limits": { 00:16:29.791 "rw_ios_per_sec": 0, 00:16:29.791 "rw_mbytes_per_sec": 0, 00:16:29.791 "r_mbytes_per_sec": 0, 00:16:29.791 "w_mbytes_per_sec": 0 00:16:29.791 }, 00:16:29.791 "claimed": false, 00:16:29.791 "zoned": false, 00:16:29.791 "supported_io_types": { 00:16:29.791 "read": true, 00:16:29.791 "write": true, 00:16:29.791 "unmap": true, 00:16:29.791 "flush": true, 00:16:29.791 "reset": true, 00:16:29.791 "nvme_admin": false, 00:16:29.791 "nvme_io": false, 00:16:29.791 "nvme_io_md": false, 00:16:29.791 "write_zeroes": true, 00:16:29.791 "zcopy": true, 00:16:29.791 "get_zone_info": false, 00:16:29.791 "zone_management": false, 00:16:29.791 "zone_append": false, 00:16:29.791 "compare": false, 00:16:29.791 "compare_and_write": false, 00:16:29.791 "abort": true, 00:16:29.791 "seek_hole": false, 00:16:29.791 "seek_data": false, 00:16:29.791 "copy": true, 00:16:29.791 "nvme_iov_md": false 00:16:29.791 }, 00:16:29.791 "memory_domains": [ 00:16:29.791 { 00:16:29.791 "dma_device_id": "system", 00:16:29.791 "dma_device_type": 1 00:16:29.791 }, 00:16:29.791 { 00:16:29.791 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:29.791 "dma_device_type": 2 00:16:29.791 } 00:16:29.791 ], 00:16:29.791 "driver_specific": { 00:16:29.791 "passthru": { 00:16:29.791 "name": "Passthru0", 00:16:29.791 "base_bdev_name": "Malloc0" 00:16:29.791 } 00:16:29.791 } 00:16:29.791 } 00:16:29.791 ]' 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.791 20:15:24 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.791 20:15:24 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:29.791 20:15:25 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.791 20:15:25 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:16:29.791 20:15:25 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.791 20:15:25 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:29.791 20:15:25 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.792 20:15:25 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:16:29.792 20:15:25 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:16:30.050 20:15:25 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:16:30.050 00:16:30.050 real 0m0.351s 00:16:30.050 user 0m0.221s 00:16:30.050 sys 0m0.039s 00:16:30.050 20:15:25 rpc.rpc_integrity -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:30.050 20:15:25 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:30.050 ************************************ 00:16:30.050 END TEST rpc_integrity 00:16:30.050 ************************************ 00:16:30.050 20:15:25 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:16:30.050 20:15:25 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:30.050 20:15:25 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:30.050 20:15:25 rpc -- common/autotest_common.sh@10 -- # set +x 00:16:30.050 ************************************ 00:16:30.050 START TEST rpc_plugins 00:16:30.050 ************************************ 00:16:30.050 20:15:25 rpc.rpc_plugins -- common/autotest_common.sh@1125 -- # rpc_plugins 00:16:30.050 20:15:25 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:16:30.050 20:15:25 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.050 20:15:25 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:16:30.050 20:15:25 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.050 20:15:25 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:16:30.050 20:15:25 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:16:30.050 20:15:25 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.050 20:15:25 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:16:30.050 20:15:25 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.050 20:15:25 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:16:30.050 { 00:16:30.050 "name": "Malloc1", 00:16:30.050 "aliases": [ 00:16:30.050 "d8e7a693-d2a7-4713-9308-d08be1d20e6c" 00:16:30.050 ], 00:16:30.050 "product_name": "Malloc disk", 00:16:30.050 "block_size": 4096, 00:16:30.050 "num_blocks": 256, 00:16:30.050 "uuid": "d8e7a693-d2a7-4713-9308-d08be1d20e6c", 00:16:30.050 "assigned_rate_limits": { 00:16:30.050 "rw_ios_per_sec": 0, 00:16:30.050 "rw_mbytes_per_sec": 0, 00:16:30.050 "r_mbytes_per_sec": 0, 00:16:30.050 "w_mbytes_per_sec": 0 00:16:30.050 }, 00:16:30.050 "claimed": false, 00:16:30.050 "zoned": false, 00:16:30.050 "supported_io_types": { 00:16:30.050 "read": true, 00:16:30.050 "write": true, 00:16:30.050 "unmap": true, 00:16:30.050 "flush": true, 00:16:30.050 "reset": true, 00:16:30.050 "nvme_admin": false, 00:16:30.050 "nvme_io": false, 00:16:30.050 "nvme_io_md": false, 00:16:30.050 "write_zeroes": true, 00:16:30.050 "zcopy": true, 00:16:30.050 "get_zone_info": false, 00:16:30.050 "zone_management": false, 00:16:30.050 "zone_append": false, 00:16:30.050 "compare": false, 00:16:30.050 "compare_and_write": false, 00:16:30.050 "abort": true, 00:16:30.050 "seek_hole": false, 00:16:30.050 "seek_data": false, 00:16:30.050 "copy": true, 00:16:30.050 "nvme_iov_md": false 00:16:30.050 }, 00:16:30.050 "memory_domains": [ 00:16:30.050 { 00:16:30.050 "dma_device_id": "system", 00:16:30.050 "dma_device_type": 1 00:16:30.050 }, 00:16:30.050 { 00:16:30.050 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:30.050 "dma_device_type": 2 00:16:30.050 } 00:16:30.050 ], 00:16:30.050 "driver_specific": {} 00:16:30.050 } 00:16:30.050 ]' 00:16:30.050 20:15:25 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:16:30.050 20:15:25 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:16:30.050 20:15:25 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:16:30.050 20:15:25 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.050 20:15:25 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:16:30.050 20:15:25 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.050 20:15:25 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:16:30.050 20:15:25 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.050 20:15:25 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:16:30.050 20:15:25 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.050 20:15:25 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:16:30.050 20:15:25 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:16:30.310 20:15:25 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:16:30.310 00:16:30.310 real 0m0.173s 00:16:30.310 user 0m0.102s 00:16:30.310 sys 0m0.029s 00:16:30.310 20:15:25 rpc.rpc_plugins -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:30.310 20:15:25 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:16:30.310 ************************************ 00:16:30.310 END TEST rpc_plugins 00:16:30.310 ************************************ 00:16:30.310 20:15:25 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:16:30.310 20:15:25 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:30.310 20:15:25 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:30.310 20:15:25 rpc -- common/autotest_common.sh@10 -- # set +x 00:16:30.310 ************************************ 00:16:30.310 START TEST rpc_trace_cmd_test 00:16:30.310 ************************************ 00:16:30.310 20:15:25 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1125 -- # rpc_trace_cmd_test 00:16:30.310 20:15:25 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:16:30.310 20:15:25 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:16:30.310 20:15:25 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.310 20:15:25 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:16:30.310 20:15:25 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.310 20:15:25 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:16:30.310 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid56831", 00:16:30.310 "tpoint_group_mask": "0x8", 00:16:30.310 "iscsi_conn": { 00:16:30.310 "mask": "0x2", 00:16:30.310 "tpoint_mask": "0x0" 00:16:30.310 }, 00:16:30.310 "scsi": { 00:16:30.310 "mask": "0x4", 00:16:30.310 "tpoint_mask": "0x0" 00:16:30.310 }, 00:16:30.310 "bdev": { 00:16:30.310 "mask": "0x8", 00:16:30.310 "tpoint_mask": "0xffffffffffffffff" 00:16:30.310 }, 00:16:30.310 "nvmf_rdma": { 00:16:30.310 "mask": "0x10", 00:16:30.310 "tpoint_mask": "0x0" 00:16:30.310 }, 00:16:30.310 "nvmf_tcp": { 00:16:30.310 "mask": "0x20", 00:16:30.310 "tpoint_mask": "0x0" 00:16:30.310 }, 00:16:30.310 "ftl": { 00:16:30.310 "mask": "0x40", 00:16:30.310 "tpoint_mask": "0x0" 00:16:30.310 }, 00:16:30.310 "blobfs": { 00:16:30.310 "mask": "0x80", 00:16:30.310 "tpoint_mask": "0x0" 00:16:30.310 }, 00:16:30.310 "dsa": { 00:16:30.310 "mask": "0x200", 00:16:30.310 "tpoint_mask": "0x0" 00:16:30.310 }, 00:16:30.310 "thread": { 00:16:30.310 "mask": "0x400", 00:16:30.310 "tpoint_mask": "0x0" 00:16:30.311 }, 00:16:30.311 "nvme_pcie": { 00:16:30.311 "mask": "0x800", 00:16:30.311 "tpoint_mask": "0x0" 00:16:30.311 }, 00:16:30.311 "iaa": { 00:16:30.311 "mask": "0x1000", 00:16:30.311 "tpoint_mask": "0x0" 00:16:30.311 }, 00:16:30.311 "nvme_tcp": { 00:16:30.311 "mask": "0x2000", 00:16:30.311 "tpoint_mask": "0x0" 00:16:30.311 }, 00:16:30.311 "bdev_nvme": { 00:16:30.311 "mask": "0x4000", 00:16:30.311 "tpoint_mask": "0x0" 00:16:30.311 }, 00:16:30.311 "sock": { 00:16:30.311 "mask": "0x8000", 00:16:30.311 "tpoint_mask": "0x0" 00:16:30.311 }, 00:16:30.311 "blob": { 00:16:30.311 "mask": "0x10000", 00:16:30.311 "tpoint_mask": "0x0" 00:16:30.311 }, 00:16:30.311 "bdev_raid": { 00:16:30.311 "mask": "0x20000", 00:16:30.311 "tpoint_mask": "0x0" 00:16:30.311 } 00:16:30.311 }' 00:16:30.311 20:15:25 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:16:30.311 20:15:25 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 18 -gt 2 ']' 00:16:30.311 20:15:25 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:16:30.311 20:15:25 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:16:30.311 20:15:25 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:16:30.311 20:15:25 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:16:30.311 20:15:25 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:16:30.569 20:15:25 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:16:30.569 20:15:25 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:16:30.569 20:15:25 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:16:30.569 00:16:30.569 real 0m0.280s 00:16:30.569 user 0m0.241s 00:16:30.569 sys 0m0.027s 00:16:30.569 20:15:25 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:30.569 20:15:25 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:16:30.569 ************************************ 00:16:30.569 END TEST rpc_trace_cmd_test 00:16:30.569 ************************************ 00:16:30.569 20:15:25 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:16:30.569 20:15:25 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:16:30.569 20:15:25 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:16:30.569 20:15:25 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:30.569 20:15:25 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:30.569 20:15:25 rpc -- common/autotest_common.sh@10 -- # set +x 00:16:30.569 ************************************ 00:16:30.569 START TEST rpc_daemon_integrity 00:16:30.569 ************************************ 00:16:30.569 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1125 -- # rpc_integrity 00:16:30.569 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:16:30.569 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.569 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:30.569 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.569 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:16:30.569 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:16:30.570 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:16:30.570 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:16:30.570 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.570 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:30.570 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.570 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:16:30.570 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:16:30.570 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.570 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:30.570 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.570 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:16:30.570 { 00:16:30.570 "name": "Malloc2", 00:16:30.570 "aliases": [ 00:16:30.570 "20a3896c-e49e-4946-b649-f1090cf49778" 00:16:30.570 ], 00:16:30.570 "product_name": "Malloc disk", 00:16:30.570 "block_size": 512, 00:16:30.570 "num_blocks": 16384, 00:16:30.570 "uuid": "20a3896c-e49e-4946-b649-f1090cf49778", 00:16:30.570 "assigned_rate_limits": { 00:16:30.570 "rw_ios_per_sec": 0, 00:16:30.570 "rw_mbytes_per_sec": 0, 00:16:30.570 "r_mbytes_per_sec": 0, 00:16:30.570 "w_mbytes_per_sec": 0 00:16:30.570 }, 00:16:30.570 "claimed": false, 00:16:30.570 "zoned": false, 00:16:30.570 "supported_io_types": { 00:16:30.570 "read": true, 00:16:30.570 "write": true, 00:16:30.570 "unmap": true, 00:16:30.570 "flush": true, 00:16:30.570 "reset": true, 00:16:30.570 "nvme_admin": false, 00:16:30.570 "nvme_io": false, 00:16:30.570 "nvme_io_md": false, 00:16:30.570 "write_zeroes": true, 00:16:30.570 "zcopy": true, 00:16:30.570 "get_zone_info": false, 00:16:30.570 "zone_management": false, 00:16:30.570 "zone_append": false, 00:16:30.570 "compare": false, 00:16:30.570 "compare_and_write": false, 00:16:30.570 "abort": true, 00:16:30.570 "seek_hole": false, 00:16:30.570 "seek_data": false, 00:16:30.570 "copy": true, 00:16:30.570 "nvme_iov_md": false 00:16:30.570 }, 00:16:30.570 "memory_domains": [ 00:16:30.570 { 00:16:30.570 "dma_device_id": "system", 00:16:30.570 "dma_device_type": 1 00:16:30.570 }, 00:16:30.570 { 00:16:30.570 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:30.570 "dma_device_type": 2 00:16:30.570 } 00:16:30.570 ], 00:16:30.570 "driver_specific": {} 00:16:30.570 } 00:16:30.570 ]' 00:16:30.570 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:16:30.829 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:16:30.829 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:16:30.829 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.829 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:30.829 [2024-10-01 20:15:25.854453] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:16:30.829 [2024-10-01 20:15:25.854534] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:30.829 [2024-10-01 20:15:25.854569] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:16:30.829 [2024-10-01 20:15:25.854590] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:30.829 [2024-10-01 20:15:25.857790] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:30.829 [2024-10-01 20:15:25.857840] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:16:30.829 Passthru0 00:16:30.829 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.829 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:16:30.829 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.829 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:30.829 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.829 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:16:30.829 { 00:16:30.829 "name": "Malloc2", 00:16:30.829 "aliases": [ 00:16:30.829 "20a3896c-e49e-4946-b649-f1090cf49778" 00:16:30.829 ], 00:16:30.829 "product_name": "Malloc disk", 00:16:30.829 "block_size": 512, 00:16:30.829 "num_blocks": 16384, 00:16:30.829 "uuid": "20a3896c-e49e-4946-b649-f1090cf49778", 00:16:30.829 "assigned_rate_limits": { 00:16:30.829 "rw_ios_per_sec": 0, 00:16:30.829 "rw_mbytes_per_sec": 0, 00:16:30.829 "r_mbytes_per_sec": 0, 00:16:30.829 "w_mbytes_per_sec": 0 00:16:30.829 }, 00:16:30.829 "claimed": true, 00:16:30.829 "claim_type": "exclusive_write", 00:16:30.829 "zoned": false, 00:16:30.829 "supported_io_types": { 00:16:30.829 "read": true, 00:16:30.829 "write": true, 00:16:30.829 "unmap": true, 00:16:30.829 "flush": true, 00:16:30.829 "reset": true, 00:16:30.829 "nvme_admin": false, 00:16:30.829 "nvme_io": false, 00:16:30.829 "nvme_io_md": false, 00:16:30.829 "write_zeroes": true, 00:16:30.829 "zcopy": true, 00:16:30.829 "get_zone_info": false, 00:16:30.829 "zone_management": false, 00:16:30.829 "zone_append": false, 00:16:30.829 "compare": false, 00:16:30.829 "compare_and_write": false, 00:16:30.829 "abort": true, 00:16:30.829 "seek_hole": false, 00:16:30.829 "seek_data": false, 00:16:30.829 "copy": true, 00:16:30.829 "nvme_iov_md": false 00:16:30.829 }, 00:16:30.829 "memory_domains": [ 00:16:30.829 { 00:16:30.829 "dma_device_id": "system", 00:16:30.830 "dma_device_type": 1 00:16:30.830 }, 00:16:30.830 { 00:16:30.830 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:30.830 "dma_device_type": 2 00:16:30.830 } 00:16:30.830 ], 00:16:30.830 "driver_specific": {} 00:16:30.830 }, 00:16:30.830 { 00:16:30.830 "name": "Passthru0", 00:16:30.830 "aliases": [ 00:16:30.830 "b89c93a8-358f-52d2-b6c7-1ec349a23ccf" 00:16:30.830 ], 00:16:30.830 "product_name": "passthru", 00:16:30.830 "block_size": 512, 00:16:30.830 "num_blocks": 16384, 00:16:30.830 "uuid": "b89c93a8-358f-52d2-b6c7-1ec349a23ccf", 00:16:30.830 "assigned_rate_limits": { 00:16:30.830 "rw_ios_per_sec": 0, 00:16:30.830 "rw_mbytes_per_sec": 0, 00:16:30.830 "r_mbytes_per_sec": 0, 00:16:30.830 "w_mbytes_per_sec": 0 00:16:30.830 }, 00:16:30.830 "claimed": false, 00:16:30.830 "zoned": false, 00:16:30.830 "supported_io_types": { 00:16:30.830 "read": true, 00:16:30.830 "write": true, 00:16:30.830 "unmap": true, 00:16:30.830 "flush": true, 00:16:30.830 "reset": true, 00:16:30.830 "nvme_admin": false, 00:16:30.830 "nvme_io": false, 00:16:30.830 "nvme_io_md": false, 00:16:30.830 "write_zeroes": true, 00:16:30.830 "zcopy": true, 00:16:30.830 "get_zone_info": false, 00:16:30.830 "zone_management": false, 00:16:30.830 "zone_append": false, 00:16:30.830 "compare": false, 00:16:30.830 "compare_and_write": false, 00:16:30.830 "abort": true, 00:16:30.830 "seek_hole": false, 00:16:30.830 "seek_data": false, 00:16:30.830 "copy": true, 00:16:30.830 "nvme_iov_md": false 00:16:30.830 }, 00:16:30.830 "memory_domains": [ 00:16:30.830 { 00:16:30.830 "dma_device_id": "system", 00:16:30.830 "dma_device_type": 1 00:16:30.830 }, 00:16:30.830 { 00:16:30.830 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:30.830 "dma_device_type": 2 00:16:30.830 } 00:16:30.830 ], 00:16:30.830 "driver_specific": { 00:16:30.830 "passthru": { 00:16:30.830 "name": "Passthru0", 00:16:30.830 "base_bdev_name": "Malloc2" 00:16:30.830 } 00:16:30.830 } 00:16:30.830 } 00:16:30.830 ]' 00:16:30.830 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:16:30.830 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:16:30.830 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:16:30.830 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.830 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:30.830 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.830 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:16:30.830 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.830 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:30.830 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.830 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:16:30.830 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.830 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:30.830 20:15:25 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.830 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:16:30.830 20:15:25 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:16:30.830 20:15:26 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:16:30.830 00:16:30.830 real 0m0.353s 00:16:30.830 user 0m0.214s 00:16:30.830 sys 0m0.046s 00:16:30.830 20:15:26 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:30.830 ************************************ 00:16:30.830 20:15:26 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:16:30.830 END TEST rpc_daemon_integrity 00:16:30.830 ************************************ 00:16:31.090 20:15:26 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:16:31.090 20:15:26 rpc -- rpc/rpc.sh@84 -- # killprocess 56831 00:16:31.090 20:15:26 rpc -- common/autotest_common.sh@950 -- # '[' -z 56831 ']' 00:16:31.090 20:15:26 rpc -- common/autotest_common.sh@954 -- # kill -0 56831 00:16:31.090 20:15:26 rpc -- common/autotest_common.sh@955 -- # uname 00:16:31.090 20:15:26 rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:31.090 20:15:26 rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 56831 00:16:31.090 20:15:26 rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:31.090 20:15:26 rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:31.090 killing process with pid 56831 00:16:31.090 20:15:26 rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 56831' 00:16:31.090 20:15:26 rpc -- common/autotest_common.sh@969 -- # kill 56831 00:16:31.090 20:15:26 rpc -- common/autotest_common.sh@974 -- # wait 56831 00:16:34.377 00:16:34.377 real 0m6.323s 00:16:34.377 user 0m6.895s 00:16:34.377 sys 0m1.078s 00:16:34.377 20:15:29 rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:34.377 20:15:29 rpc -- common/autotest_common.sh@10 -- # set +x 00:16:34.377 ************************************ 00:16:34.377 END TEST rpc 00:16:34.377 ************************************ 00:16:34.377 20:15:29 -- spdk/autotest.sh@157 -- # run_test skip_rpc /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:16:34.377 20:15:29 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:34.377 20:15:29 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:34.377 20:15:29 -- common/autotest_common.sh@10 -- # set +x 00:16:34.377 ************************************ 00:16:34.377 START TEST skip_rpc 00:16:34.377 ************************************ 00:16:34.377 20:15:29 skip_rpc -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:16:34.377 * Looking for test storage... 00:16:34.377 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:16:34.377 20:15:29 skip_rpc -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:16:34.377 20:15:29 skip_rpc -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:16:34.377 20:15:29 skip_rpc -- common/autotest_common.sh@1681 -- # lcov --version 00:16:34.377 20:15:29 skip_rpc -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@344 -- # case "$op" in 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@345 -- # : 1 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@365 -- # decimal 1 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@353 -- # local d=1 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@355 -- # echo 1 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@366 -- # decimal 2 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@353 -- # local d=2 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@355 -- # echo 2 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:34.377 20:15:29 skip_rpc -- scripts/common.sh@368 -- # return 0 00:16:34.377 20:15:29 skip_rpc -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:34.377 20:15:29 skip_rpc -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:16:34.377 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:34.377 --rc genhtml_branch_coverage=1 00:16:34.377 --rc genhtml_function_coverage=1 00:16:34.377 --rc genhtml_legend=1 00:16:34.377 --rc geninfo_all_blocks=1 00:16:34.377 --rc geninfo_unexecuted_blocks=1 00:16:34.377 00:16:34.377 ' 00:16:34.377 20:15:29 skip_rpc -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:16:34.377 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:34.377 --rc genhtml_branch_coverage=1 00:16:34.377 --rc genhtml_function_coverage=1 00:16:34.377 --rc genhtml_legend=1 00:16:34.377 --rc geninfo_all_blocks=1 00:16:34.377 --rc geninfo_unexecuted_blocks=1 00:16:34.377 00:16:34.377 ' 00:16:34.377 20:15:29 skip_rpc -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:16:34.377 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:34.377 --rc genhtml_branch_coverage=1 00:16:34.377 --rc genhtml_function_coverage=1 00:16:34.377 --rc genhtml_legend=1 00:16:34.377 --rc geninfo_all_blocks=1 00:16:34.377 --rc geninfo_unexecuted_blocks=1 00:16:34.377 00:16:34.377 ' 00:16:34.377 20:15:29 skip_rpc -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:16:34.377 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:34.377 --rc genhtml_branch_coverage=1 00:16:34.377 --rc genhtml_function_coverage=1 00:16:34.377 --rc genhtml_legend=1 00:16:34.377 --rc geninfo_all_blocks=1 00:16:34.377 --rc geninfo_unexecuted_blocks=1 00:16:34.377 00:16:34.377 ' 00:16:34.377 20:15:29 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:16:34.377 20:15:29 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:16:34.377 20:15:29 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:16:34.377 20:15:29 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:34.377 20:15:29 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:34.377 20:15:29 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:34.377 ************************************ 00:16:34.377 START TEST skip_rpc 00:16:34.377 ************************************ 00:16:34.377 20:15:29 skip_rpc.skip_rpc -- common/autotest_common.sh@1125 -- # test_skip_rpc 00:16:34.377 20:15:29 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=57072 00:16:34.377 20:15:29 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:16:34.377 20:15:29 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:16:34.377 20:15:29 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:16:34.377 [2024-10-01 20:15:29.453736] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:16:34.377 [2024-10-01 20:15:29.453945] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57072 ] 00:16:34.636 [2024-10-01 20:15:29.638817] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:34.895 [2024-10-01 20:15:29.949936] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@650 -- # local es=0 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd spdk_get_version 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@653 -- # rpc_cmd spdk_get_version 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@653 -- # es=1 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 57072 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@950 -- # '[' -z 57072 ']' 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # kill -0 57072 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@955 -- # uname 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:39.081 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 57072 00:16:39.340 killing process with pid 57072 00:16:39.340 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:39.340 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:39.340 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 57072' 00:16:39.340 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@969 -- # kill 57072 00:16:39.340 20:15:34 skip_rpc.skip_rpc -- common/autotest_common.sh@974 -- # wait 57072 00:16:42.633 00:16:42.633 real 0m8.054s 00:16:42.633 user 0m7.356s 00:16:42.633 sys 0m0.581s 00:16:42.633 20:15:37 skip_rpc.skip_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:42.633 ************************************ 00:16:42.633 END TEST skip_rpc 00:16:42.633 ************************************ 00:16:42.633 20:15:37 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:42.633 20:15:37 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:16:42.633 20:15:37 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:42.633 20:15:37 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:42.633 20:15:37 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:42.633 ************************************ 00:16:42.633 START TEST skip_rpc_with_json 00:16:42.633 ************************************ 00:16:42.633 20:15:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1125 -- # test_skip_rpc_with_json 00:16:42.633 20:15:37 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:16:42.633 20:15:37 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=57186 00:16:42.633 20:15:37 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:16:42.633 20:15:37 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 57186 00:16:42.633 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:42.633 20:15:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@831 -- # '[' -z 57186 ']' 00:16:42.633 20:15:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:42.633 20:15:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:42.633 20:15:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:42.633 20:15:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:42.633 20:15:37 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:16:42.633 20:15:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:16:42.633 [2024-10-01 20:15:37.563244] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:16:42.633 [2024-10-01 20:15:37.564004] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57186 ] 00:16:42.633 [2024-10-01 20:15:37.742876] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:42.891 [2024-10-01 20:15:37.974528] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:44.325 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:44.325 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@864 -- # return 0 00:16:44.325 20:15:39 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:16:44.325 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:44.325 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:16:44.325 [2024-10-01 20:15:39.102815] nvmf_rpc.c:2703:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:16:44.325 request: 00:16:44.325 { 00:16:44.325 "trtype": "tcp", 00:16:44.325 "method": "nvmf_get_transports", 00:16:44.325 "req_id": 1 00:16:44.325 } 00:16:44.325 Got JSON-RPC error response 00:16:44.325 response: 00:16:44.325 { 00:16:44.325 "code": -19, 00:16:44.325 "message": "No such device" 00:16:44.325 } 00:16:44.325 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:44.325 20:15:39 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:16:44.325 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:44.325 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:16:44.325 [2024-10-01 20:15:39.114990] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:16:44.325 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:44.325 20:15:39 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:16:44.325 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:44.325 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:16:44.325 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:44.325 20:15:39 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:16:44.325 { 00:16:44.325 "subsystems": [ 00:16:44.325 { 00:16:44.325 "subsystem": "fsdev", 00:16:44.325 "config": [ 00:16:44.325 { 00:16:44.325 "method": "fsdev_set_opts", 00:16:44.325 "params": { 00:16:44.325 "fsdev_io_pool_size": 65535, 00:16:44.325 "fsdev_io_cache_size": 256 00:16:44.325 } 00:16:44.325 } 00:16:44.325 ] 00:16:44.325 }, 00:16:44.325 { 00:16:44.325 "subsystem": "keyring", 00:16:44.325 "config": [] 00:16:44.325 }, 00:16:44.325 { 00:16:44.325 "subsystem": "iobuf", 00:16:44.325 "config": [ 00:16:44.325 { 00:16:44.325 "method": "iobuf_set_options", 00:16:44.325 "params": { 00:16:44.325 "small_pool_count": 8192, 00:16:44.325 "large_pool_count": 1024, 00:16:44.325 "small_bufsize": 8192, 00:16:44.325 "large_bufsize": 135168 00:16:44.325 } 00:16:44.325 } 00:16:44.325 ] 00:16:44.325 }, 00:16:44.325 { 00:16:44.325 "subsystem": "sock", 00:16:44.325 "config": [ 00:16:44.325 { 00:16:44.325 "method": "sock_set_default_impl", 00:16:44.325 "params": { 00:16:44.325 "impl_name": "posix" 00:16:44.325 } 00:16:44.325 }, 00:16:44.325 { 00:16:44.325 "method": "sock_impl_set_options", 00:16:44.325 "params": { 00:16:44.325 "impl_name": "ssl", 00:16:44.325 "recv_buf_size": 4096, 00:16:44.325 "send_buf_size": 4096, 00:16:44.325 "enable_recv_pipe": true, 00:16:44.325 "enable_quickack": false, 00:16:44.325 "enable_placement_id": 0, 00:16:44.325 "enable_zerocopy_send_server": true, 00:16:44.325 "enable_zerocopy_send_client": false, 00:16:44.325 "zerocopy_threshold": 0, 00:16:44.325 "tls_version": 0, 00:16:44.325 "enable_ktls": false 00:16:44.325 } 00:16:44.325 }, 00:16:44.325 { 00:16:44.325 "method": "sock_impl_set_options", 00:16:44.325 "params": { 00:16:44.325 "impl_name": "posix", 00:16:44.325 "recv_buf_size": 2097152, 00:16:44.325 "send_buf_size": 2097152, 00:16:44.325 "enable_recv_pipe": true, 00:16:44.326 "enable_quickack": false, 00:16:44.326 "enable_placement_id": 0, 00:16:44.326 "enable_zerocopy_send_server": true, 00:16:44.326 "enable_zerocopy_send_client": false, 00:16:44.326 "zerocopy_threshold": 0, 00:16:44.326 "tls_version": 0, 00:16:44.326 "enable_ktls": false 00:16:44.326 } 00:16:44.326 } 00:16:44.326 ] 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "subsystem": "vmd", 00:16:44.326 "config": [] 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "subsystem": "accel", 00:16:44.326 "config": [ 00:16:44.326 { 00:16:44.326 "method": "accel_set_options", 00:16:44.326 "params": { 00:16:44.326 "small_cache_size": 128, 00:16:44.326 "large_cache_size": 16, 00:16:44.326 "task_count": 2048, 00:16:44.326 "sequence_count": 2048, 00:16:44.326 "buf_count": 2048 00:16:44.326 } 00:16:44.326 } 00:16:44.326 ] 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "subsystem": "bdev", 00:16:44.326 "config": [ 00:16:44.326 { 00:16:44.326 "method": "bdev_set_options", 00:16:44.326 "params": { 00:16:44.326 "bdev_io_pool_size": 65535, 00:16:44.326 "bdev_io_cache_size": 256, 00:16:44.326 "bdev_auto_examine": true, 00:16:44.326 "iobuf_small_cache_size": 128, 00:16:44.326 "iobuf_large_cache_size": 16, 00:16:44.326 "bdev_io_stack_size": 4096 00:16:44.326 } 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "method": "bdev_raid_set_options", 00:16:44.326 "params": { 00:16:44.326 "process_window_size_kb": 1024, 00:16:44.326 "process_max_bandwidth_mb_sec": 0 00:16:44.326 } 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "method": "bdev_iscsi_set_options", 00:16:44.326 "params": { 00:16:44.326 "timeout_sec": 30 00:16:44.326 } 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "method": "bdev_nvme_set_options", 00:16:44.326 "params": { 00:16:44.326 "action_on_timeout": "none", 00:16:44.326 "timeout_us": 0, 00:16:44.326 "timeout_admin_us": 0, 00:16:44.326 "keep_alive_timeout_ms": 10000, 00:16:44.326 "arbitration_burst": 0, 00:16:44.326 "low_priority_weight": 0, 00:16:44.326 "medium_priority_weight": 0, 00:16:44.326 "high_priority_weight": 0, 00:16:44.326 "nvme_adminq_poll_period_us": 10000, 00:16:44.326 "nvme_ioq_poll_period_us": 0, 00:16:44.326 "io_queue_requests": 0, 00:16:44.326 "delay_cmd_submit": true, 00:16:44.326 "transport_retry_count": 4, 00:16:44.326 "bdev_retry_count": 3, 00:16:44.326 "transport_ack_timeout": 0, 00:16:44.326 "ctrlr_loss_timeout_sec": 0, 00:16:44.326 "reconnect_delay_sec": 0, 00:16:44.326 "fast_io_fail_timeout_sec": 0, 00:16:44.326 "disable_auto_failback": false, 00:16:44.326 "generate_uuids": false, 00:16:44.326 "transport_tos": 0, 00:16:44.326 "nvme_error_stat": false, 00:16:44.326 "rdma_srq_size": 0, 00:16:44.326 "io_path_stat": false, 00:16:44.326 "allow_accel_sequence": false, 00:16:44.326 "rdma_max_cq_size": 0, 00:16:44.326 "rdma_cm_event_timeout_ms": 0, 00:16:44.326 "dhchap_digests": [ 00:16:44.326 "sha256", 00:16:44.326 "sha384", 00:16:44.326 "sha512" 00:16:44.326 ], 00:16:44.326 "dhchap_dhgroups": [ 00:16:44.326 "null", 00:16:44.326 "ffdhe2048", 00:16:44.326 "ffdhe3072", 00:16:44.326 "ffdhe4096", 00:16:44.326 "ffdhe6144", 00:16:44.326 "ffdhe8192" 00:16:44.326 ] 00:16:44.326 } 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "method": "bdev_nvme_set_hotplug", 00:16:44.326 "params": { 00:16:44.326 "period_us": 100000, 00:16:44.326 "enable": false 00:16:44.326 } 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "method": "bdev_wait_for_examine" 00:16:44.326 } 00:16:44.326 ] 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "subsystem": "scsi", 00:16:44.326 "config": null 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "subsystem": "scheduler", 00:16:44.326 "config": [ 00:16:44.326 { 00:16:44.326 "method": "framework_set_scheduler", 00:16:44.326 "params": { 00:16:44.326 "name": "static" 00:16:44.326 } 00:16:44.326 } 00:16:44.326 ] 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "subsystem": "vhost_scsi", 00:16:44.326 "config": [] 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "subsystem": "vhost_blk", 00:16:44.326 "config": [] 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "subsystem": "ublk", 00:16:44.326 "config": [] 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "subsystem": "nbd", 00:16:44.326 "config": [] 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "subsystem": "nvmf", 00:16:44.326 "config": [ 00:16:44.326 { 00:16:44.326 "method": "nvmf_set_config", 00:16:44.326 "params": { 00:16:44.326 "discovery_filter": "match_any", 00:16:44.326 "admin_cmd_passthru": { 00:16:44.326 "identify_ctrlr": false 00:16:44.326 }, 00:16:44.326 "dhchap_digests": [ 00:16:44.326 "sha256", 00:16:44.326 "sha384", 00:16:44.326 "sha512" 00:16:44.326 ], 00:16:44.326 "dhchap_dhgroups": [ 00:16:44.326 "null", 00:16:44.326 "ffdhe2048", 00:16:44.326 "ffdhe3072", 00:16:44.326 "ffdhe4096", 00:16:44.326 "ffdhe6144", 00:16:44.326 "ffdhe8192" 00:16:44.326 ] 00:16:44.326 } 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "method": "nvmf_set_max_subsystems", 00:16:44.326 "params": { 00:16:44.326 "max_subsystems": 1024 00:16:44.326 } 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "method": "nvmf_set_crdt", 00:16:44.326 "params": { 00:16:44.326 "crdt1": 0, 00:16:44.326 "crdt2": 0, 00:16:44.326 "crdt3": 0 00:16:44.326 } 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "method": "nvmf_create_transport", 00:16:44.326 "params": { 00:16:44.326 "trtype": "TCP", 00:16:44.326 "max_queue_depth": 128, 00:16:44.326 "max_io_qpairs_per_ctrlr": 127, 00:16:44.326 "in_capsule_data_size": 4096, 00:16:44.326 "max_io_size": 131072, 00:16:44.326 "io_unit_size": 131072, 00:16:44.326 "max_aq_depth": 128, 00:16:44.326 "num_shared_buffers": 511, 00:16:44.326 "buf_cache_size": 4294967295, 00:16:44.326 "dif_insert_or_strip": false, 00:16:44.326 "zcopy": false, 00:16:44.326 "c2h_success": true, 00:16:44.326 "sock_priority": 0, 00:16:44.326 "abort_timeout_sec": 1, 00:16:44.326 "ack_timeout": 0, 00:16:44.326 "data_wr_pool_size": 0 00:16:44.326 } 00:16:44.326 } 00:16:44.326 ] 00:16:44.326 }, 00:16:44.326 { 00:16:44.326 "subsystem": "iscsi", 00:16:44.326 "config": [ 00:16:44.326 { 00:16:44.326 "method": "iscsi_set_options", 00:16:44.326 "params": { 00:16:44.326 "node_base": "iqn.2016-06.io.spdk", 00:16:44.326 "max_sessions": 128, 00:16:44.326 "max_connections_per_session": 2, 00:16:44.326 "max_queue_depth": 64, 00:16:44.326 "default_time2wait": 2, 00:16:44.326 "default_time2retain": 20, 00:16:44.326 "first_burst_length": 8192, 00:16:44.326 "immediate_data": true, 00:16:44.326 "allow_duplicated_isid": false, 00:16:44.326 "error_recovery_level": 0, 00:16:44.326 "nop_timeout": 60, 00:16:44.326 "nop_in_interval": 30, 00:16:44.326 "disable_chap": false, 00:16:44.326 "require_chap": false, 00:16:44.326 "mutual_chap": false, 00:16:44.326 "chap_group": 0, 00:16:44.326 "max_large_datain_per_connection": 64, 00:16:44.326 "max_r2t_per_connection": 4, 00:16:44.326 "pdu_pool_size": 36864, 00:16:44.326 "immediate_data_pool_size": 16384, 00:16:44.327 "data_out_pool_size": 2048 00:16:44.327 } 00:16:44.327 } 00:16:44.327 ] 00:16:44.327 } 00:16:44.327 ] 00:16:44.327 } 00:16:44.327 20:15:39 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:16:44.327 20:15:39 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 57186 00:16:44.327 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@950 -- # '[' -z 57186 ']' 00:16:44.327 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # kill -0 57186 00:16:44.327 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # uname 00:16:44.327 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:44.327 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 57186 00:16:44.327 killing process with pid 57186 00:16:44.327 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:44.327 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:44.327 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@968 -- # echo 'killing process with pid 57186' 00:16:44.327 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@969 -- # kill 57186 00:16:44.327 20:15:39 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@974 -- # wait 57186 00:16:47.610 20:15:42 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=57254 00:16:47.610 20:15:42 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:16:47.610 20:15:42 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:16:52.958 20:15:47 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 57254 00:16:52.958 20:15:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@950 -- # '[' -z 57254 ']' 00:16:52.958 20:15:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # kill -0 57254 00:16:52.958 20:15:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # uname 00:16:52.958 20:15:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:52.958 20:15:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 57254 00:16:52.958 killing process with pid 57254 00:16:52.958 20:15:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:52.958 20:15:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:52.958 20:15:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@968 -- # echo 'killing process with pid 57254' 00:16:52.958 20:15:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@969 -- # kill 57254 00:16:52.958 20:15:47 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@974 -- # wait 57254 00:16:55.528 20:15:50 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:16:55.528 20:15:50 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:16:55.528 ************************************ 00:16:55.528 END TEST skip_rpc_with_json 00:16:55.528 ************************************ 00:16:55.528 00:16:55.528 real 0m12.960s 00:16:55.528 user 0m12.132s 00:16:55.528 sys 0m1.256s 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:16:55.529 20:15:50 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:16:55.529 20:15:50 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:55.529 20:15:50 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:55.529 20:15:50 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:55.529 ************************************ 00:16:55.529 START TEST skip_rpc_with_delay 00:16:55.529 ************************************ 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1125 -- # test_skip_rpc_with_delay 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@650 -- # local es=0 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@652 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@638 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@653 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:16:55.529 [2024-10-01 20:15:50.555211] app.c: 840:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:16:55.529 [2024-10-01 20:15:50.555434] app.c: 719:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 0, errno: 2 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@653 -- # es=1 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:55.529 00:16:55.529 real 0m0.187s 00:16:55.529 user 0m0.091s 00:16:55.529 sys 0m0.091s 00:16:55.529 ************************************ 00:16:55.529 END TEST skip_rpc_with_delay 00:16:55.529 ************************************ 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:55.529 20:15:50 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:16:55.529 20:15:50 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:16:55.529 20:15:50 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:16:55.529 20:15:50 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:16:55.529 20:15:50 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:55.529 20:15:50 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:55.529 20:15:50 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:55.529 ************************************ 00:16:55.529 START TEST exit_on_failed_rpc_init 00:16:55.529 ************************************ 00:16:55.529 20:15:50 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1125 -- # test_exit_on_failed_rpc_init 00:16:55.529 20:15:50 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=57393 00:16:55.529 20:15:50 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:16:55.529 20:15:50 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 57393 00:16:55.529 20:15:50 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@831 -- # '[' -z 57393 ']' 00:16:55.529 20:15:50 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:55.529 20:15:50 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:55.529 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:55.529 20:15:50 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:55.529 20:15:50 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:55.529 20:15:50 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:16:55.788 [2024-10-01 20:15:50.814697] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:16:55.788 [2024-10-01 20:15:50.814901] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57393 ] 00:16:55.788 [2024-10-01 20:15:50.992228] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:56.047 [2024-10-01 20:15:51.297697] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:57.424 20:15:52 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:57.424 20:15:52 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@864 -- # return 0 00:16:57.424 20:15:52 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:16:57.424 20:15:52 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:16:57.424 20:15:52 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@650 -- # local es=0 00:16:57.424 20:15:52 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@652 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:16:57.424 20:15:52 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@638 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:16:57.424 20:15:52 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:57.424 20:15:52 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:16:57.424 20:15:52 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:57.424 20:15:52 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:16:57.424 20:15:52 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:57.424 20:15:52 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:16:57.424 20:15:52 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:16:57.424 20:15:52 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@653 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:16:57.424 [2024-10-01 20:15:52.661540] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:16:57.424 [2024-10-01 20:15:52.661705] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57417 ] 00:16:57.682 [2024-10-01 20:15:52.833080] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:57.940 [2024-10-01 20:15:53.099353] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:16:57.940 [2024-10-01 20:15:53.099773] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:16:57.940 [2024-10-01 20:15:53.099806] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:16:57.940 [2024-10-01 20:15:53.099826] app.c:1061:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@653 -- # es=234 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@662 -- # es=106 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@663 -- # case "$es" in 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@670 -- # es=1 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 57393 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@950 -- # '[' -z 57393 ']' 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # kill -0 57393 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@955 -- # uname 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 57393 00:16:58.507 killing process with pid 57393 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@968 -- # echo 'killing process with pid 57393' 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@969 -- # kill 57393 00:16:58.507 20:15:53 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@974 -- # wait 57393 00:17:01.861 ************************************ 00:17:01.861 END TEST exit_on_failed_rpc_init 00:17:01.861 ************************************ 00:17:01.861 00:17:01.861 real 0m5.902s 00:17:01.861 user 0m6.410s 00:17:01.861 sys 0m0.844s 00:17:01.861 20:15:56 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:01.861 20:15:56 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:17:01.861 20:15:56 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:17:01.861 ************************************ 00:17:01.861 END TEST skip_rpc 00:17:01.861 ************************************ 00:17:01.861 00:17:01.861 real 0m27.514s 00:17:01.861 user 0m26.164s 00:17:01.861 sys 0m3.002s 00:17:01.861 20:15:56 skip_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:01.861 20:15:56 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:17:01.861 20:15:56 -- spdk/autotest.sh@158 -- # run_test rpc_client /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:17:01.861 20:15:56 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:17:01.861 20:15:56 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:01.861 20:15:56 -- common/autotest_common.sh@10 -- # set +x 00:17:01.861 ************************************ 00:17:01.861 START TEST rpc_client 00:17:01.861 ************************************ 00:17:01.861 20:15:56 rpc_client -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:17:01.861 * Looking for test storage... 00:17:01.861 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc_client 00:17:01.861 20:15:56 rpc_client -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:17:01.861 20:15:56 rpc_client -- common/autotest_common.sh@1681 -- # lcov --version 00:17:01.861 20:15:56 rpc_client -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:17:01.861 20:15:56 rpc_client -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@336 -- # IFS=.-: 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@336 -- # read -ra ver1 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@337 -- # IFS=.-: 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@337 -- # read -ra ver2 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@338 -- # local 'op=<' 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@340 -- # ver1_l=2 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@341 -- # ver2_l=1 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@344 -- # case "$op" in 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@345 -- # : 1 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@365 -- # decimal 1 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@353 -- # local d=1 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@355 -- # echo 1 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@365 -- # ver1[v]=1 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@366 -- # decimal 2 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@353 -- # local d=2 00:17:01.861 20:15:56 rpc_client -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:01.862 20:15:56 rpc_client -- scripts/common.sh@355 -- # echo 2 00:17:01.862 20:15:56 rpc_client -- scripts/common.sh@366 -- # ver2[v]=2 00:17:01.862 20:15:56 rpc_client -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:01.862 20:15:56 rpc_client -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:01.862 20:15:56 rpc_client -- scripts/common.sh@368 -- # return 0 00:17:01.862 20:15:56 rpc_client -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:01.862 20:15:56 rpc_client -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:17:01.862 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:01.862 --rc genhtml_branch_coverage=1 00:17:01.862 --rc genhtml_function_coverage=1 00:17:01.862 --rc genhtml_legend=1 00:17:01.862 --rc geninfo_all_blocks=1 00:17:01.862 --rc geninfo_unexecuted_blocks=1 00:17:01.862 00:17:01.862 ' 00:17:01.862 20:15:56 rpc_client -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:17:01.862 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:01.862 --rc genhtml_branch_coverage=1 00:17:01.862 --rc genhtml_function_coverage=1 00:17:01.862 --rc genhtml_legend=1 00:17:01.862 --rc geninfo_all_blocks=1 00:17:01.862 --rc geninfo_unexecuted_blocks=1 00:17:01.862 00:17:01.862 ' 00:17:01.862 20:15:56 rpc_client -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:17:01.862 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:01.862 --rc genhtml_branch_coverage=1 00:17:01.862 --rc genhtml_function_coverage=1 00:17:01.862 --rc genhtml_legend=1 00:17:01.862 --rc geninfo_all_blocks=1 00:17:01.862 --rc geninfo_unexecuted_blocks=1 00:17:01.862 00:17:01.862 ' 00:17:01.862 20:15:56 rpc_client -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:17:01.862 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:01.862 --rc genhtml_branch_coverage=1 00:17:01.862 --rc genhtml_function_coverage=1 00:17:01.862 --rc genhtml_legend=1 00:17:01.862 --rc geninfo_all_blocks=1 00:17:01.862 --rc geninfo_unexecuted_blocks=1 00:17:01.862 00:17:01.862 ' 00:17:01.862 20:15:56 rpc_client -- rpc_client/rpc_client.sh@10 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client_test 00:17:01.862 OK 00:17:01.862 20:15:56 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:17:01.862 00:17:01.862 real 0m0.262s 00:17:01.862 user 0m0.149s 00:17:01.862 sys 0m0.119s 00:17:01.862 20:15:56 rpc_client -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:01.862 20:15:56 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:17:01.862 ************************************ 00:17:01.862 END TEST rpc_client 00:17:01.862 ************************************ 00:17:01.862 20:15:56 -- spdk/autotest.sh@159 -- # run_test json_config /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:17:01.862 20:15:56 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:17:01.862 20:15:56 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:01.862 20:15:56 -- common/autotest_common.sh@10 -- # set +x 00:17:01.862 ************************************ 00:17:01.862 START TEST json_config 00:17:01.862 ************************************ 00:17:01.862 20:15:56 json_config -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:17:01.862 20:15:57 json_config -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:17:01.862 20:15:57 json_config -- common/autotest_common.sh@1681 -- # lcov --version 00:17:01.862 20:15:57 json_config -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:17:02.121 20:15:57 json_config -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:17:02.121 20:15:57 json_config -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:02.121 20:15:57 json_config -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:02.121 20:15:57 json_config -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:02.121 20:15:57 json_config -- scripts/common.sh@336 -- # IFS=.-: 00:17:02.121 20:15:57 json_config -- scripts/common.sh@336 -- # read -ra ver1 00:17:02.121 20:15:57 json_config -- scripts/common.sh@337 -- # IFS=.-: 00:17:02.121 20:15:57 json_config -- scripts/common.sh@337 -- # read -ra ver2 00:17:02.121 20:15:57 json_config -- scripts/common.sh@338 -- # local 'op=<' 00:17:02.121 20:15:57 json_config -- scripts/common.sh@340 -- # ver1_l=2 00:17:02.121 20:15:57 json_config -- scripts/common.sh@341 -- # ver2_l=1 00:17:02.121 20:15:57 json_config -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:02.121 20:15:57 json_config -- scripts/common.sh@344 -- # case "$op" in 00:17:02.121 20:15:57 json_config -- scripts/common.sh@345 -- # : 1 00:17:02.121 20:15:57 json_config -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:02.121 20:15:57 json_config -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:02.121 20:15:57 json_config -- scripts/common.sh@365 -- # decimal 1 00:17:02.121 20:15:57 json_config -- scripts/common.sh@353 -- # local d=1 00:17:02.121 20:15:57 json_config -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:02.121 20:15:57 json_config -- scripts/common.sh@355 -- # echo 1 00:17:02.121 20:15:57 json_config -- scripts/common.sh@365 -- # ver1[v]=1 00:17:02.121 20:15:57 json_config -- scripts/common.sh@366 -- # decimal 2 00:17:02.121 20:15:57 json_config -- scripts/common.sh@353 -- # local d=2 00:17:02.121 20:15:57 json_config -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:02.121 20:15:57 json_config -- scripts/common.sh@355 -- # echo 2 00:17:02.121 20:15:57 json_config -- scripts/common.sh@366 -- # ver2[v]=2 00:17:02.121 20:15:57 json_config -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:02.121 20:15:57 json_config -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:02.121 20:15:57 json_config -- scripts/common.sh@368 -- # return 0 00:17:02.122 20:15:57 json_config -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:02.122 20:15:57 json_config -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:17:02.122 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:02.122 --rc genhtml_branch_coverage=1 00:17:02.122 --rc genhtml_function_coverage=1 00:17:02.122 --rc genhtml_legend=1 00:17:02.122 --rc geninfo_all_blocks=1 00:17:02.122 --rc geninfo_unexecuted_blocks=1 00:17:02.122 00:17:02.122 ' 00:17:02.122 20:15:57 json_config -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:17:02.122 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:02.122 --rc genhtml_branch_coverage=1 00:17:02.122 --rc genhtml_function_coverage=1 00:17:02.122 --rc genhtml_legend=1 00:17:02.122 --rc geninfo_all_blocks=1 00:17:02.122 --rc geninfo_unexecuted_blocks=1 00:17:02.122 00:17:02.122 ' 00:17:02.122 20:15:57 json_config -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:17:02.122 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:02.122 --rc genhtml_branch_coverage=1 00:17:02.122 --rc genhtml_function_coverage=1 00:17:02.122 --rc genhtml_legend=1 00:17:02.122 --rc geninfo_all_blocks=1 00:17:02.122 --rc geninfo_unexecuted_blocks=1 00:17:02.122 00:17:02.122 ' 00:17:02.122 20:15:57 json_config -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:17:02.122 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:02.122 --rc genhtml_branch_coverage=1 00:17:02.122 --rc genhtml_function_coverage=1 00:17:02.122 --rc genhtml_legend=1 00:17:02.122 --rc geninfo_all_blocks=1 00:17:02.122 --rc geninfo_unexecuted_blocks=1 00:17:02.122 00:17:02.122 ' 00:17:02.122 20:15:57 json_config -- json_config/json_config.sh@8 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@7 -- # uname -s 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5a683382-c549-400e-8b27-e29f159572f4 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=5a683382-c549-400e-8b27-e29f159572f4 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:17:02.122 20:15:57 json_config -- scripts/common.sh@15 -- # shopt -s extglob 00:17:02.122 20:15:57 json_config -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:17:02.122 20:15:57 json_config -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:17:02.122 20:15:57 json_config -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:17:02.122 20:15:57 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:02.122 20:15:57 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:02.122 20:15:57 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:02.122 20:15:57 json_config -- paths/export.sh@5 -- # export PATH 00:17:02.122 20:15:57 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@51 -- # : 0 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:17:02.122 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:17:02.122 20:15:57 json_config -- nvmf/common.sh@55 -- # have_pci_nics=0 00:17:02.122 20:15:57 json_config -- json_config/json_config.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:17:02.122 20:15:57 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:17:02.122 20:15:57 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:17:02.122 20:15:57 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:17:02.122 20:15:57 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:17:02.122 20:15:57 json_config -- json_config/json_config.sh@27 -- # echo 'WARNING: No tests are enabled so not running JSON configuration tests' 00:17:02.122 WARNING: No tests are enabled so not running JSON configuration tests 00:17:02.122 20:15:57 json_config -- json_config/json_config.sh@28 -- # exit 0 00:17:02.122 00:17:02.122 real 0m0.200s 00:17:02.122 user 0m0.125s 00:17:02.122 sys 0m0.075s 00:17:02.122 20:15:57 json_config -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:02.122 20:15:57 json_config -- common/autotest_common.sh@10 -- # set +x 00:17:02.122 ************************************ 00:17:02.122 END TEST json_config 00:17:02.122 ************************************ 00:17:02.122 20:15:57 -- spdk/autotest.sh@160 -- # run_test json_config_extra_key /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:17:02.122 20:15:57 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:17:02.122 20:15:57 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:02.122 20:15:57 -- common/autotest_common.sh@10 -- # set +x 00:17:02.122 ************************************ 00:17:02.122 START TEST json_config_extra_key 00:17:02.122 ************************************ 00:17:02.122 20:15:57 json_config_extra_key -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:17:02.122 20:15:57 json_config_extra_key -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:17:02.122 20:15:57 json_config_extra_key -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:17:02.122 20:15:57 json_config_extra_key -- common/autotest_common.sh@1681 -- # lcov --version 00:17:02.395 20:15:57 json_config_extra_key -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@336 -- # IFS=.-: 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@336 -- # read -ra ver1 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@337 -- # IFS=.-: 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@337 -- # read -ra ver2 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@338 -- # local 'op=<' 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@340 -- # ver1_l=2 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@341 -- # ver2_l=1 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@344 -- # case "$op" in 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@345 -- # : 1 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@365 -- # decimal 1 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@353 -- # local d=1 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@355 -- # echo 1 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@365 -- # ver1[v]=1 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@366 -- # decimal 2 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@353 -- # local d=2 00:17:02.395 20:15:57 json_config_extra_key -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:02.396 20:15:57 json_config_extra_key -- scripts/common.sh@355 -- # echo 2 00:17:02.396 20:15:57 json_config_extra_key -- scripts/common.sh@366 -- # ver2[v]=2 00:17:02.396 20:15:57 json_config_extra_key -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:02.396 20:15:57 json_config_extra_key -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:02.396 20:15:57 json_config_extra_key -- scripts/common.sh@368 -- # return 0 00:17:02.396 20:15:57 json_config_extra_key -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:02.396 20:15:57 json_config_extra_key -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:17:02.396 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:02.396 --rc genhtml_branch_coverage=1 00:17:02.396 --rc genhtml_function_coverage=1 00:17:02.396 --rc genhtml_legend=1 00:17:02.396 --rc geninfo_all_blocks=1 00:17:02.396 --rc geninfo_unexecuted_blocks=1 00:17:02.396 00:17:02.396 ' 00:17:02.396 20:15:57 json_config_extra_key -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:17:02.396 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:02.396 --rc genhtml_branch_coverage=1 00:17:02.396 --rc genhtml_function_coverage=1 00:17:02.396 --rc genhtml_legend=1 00:17:02.396 --rc geninfo_all_blocks=1 00:17:02.396 --rc geninfo_unexecuted_blocks=1 00:17:02.396 00:17:02.396 ' 00:17:02.396 20:15:57 json_config_extra_key -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:17:02.396 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:02.396 --rc genhtml_branch_coverage=1 00:17:02.396 --rc genhtml_function_coverage=1 00:17:02.396 --rc genhtml_legend=1 00:17:02.396 --rc geninfo_all_blocks=1 00:17:02.396 --rc geninfo_unexecuted_blocks=1 00:17:02.396 00:17:02.396 ' 00:17:02.396 20:15:57 json_config_extra_key -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:17:02.396 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:02.396 --rc genhtml_branch_coverage=1 00:17:02.396 --rc genhtml_function_coverage=1 00:17:02.396 --rc genhtml_legend=1 00:17:02.396 --rc geninfo_all_blocks=1 00:17:02.396 --rc geninfo_unexecuted_blocks=1 00:17:02.396 00:17:02.396 ' 00:17:02.396 20:15:57 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5a683382-c549-400e-8b27-e29f159572f4 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=5a683382-c549-400e-8b27-e29f159572f4 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:17:02.396 20:15:57 json_config_extra_key -- scripts/common.sh@15 -- # shopt -s extglob 00:17:02.396 20:15:57 json_config_extra_key -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:17:02.396 20:15:57 json_config_extra_key -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:17:02.396 20:15:57 json_config_extra_key -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:17:02.396 20:15:57 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:02.396 20:15:57 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:02.396 20:15:57 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:02.396 20:15:57 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:17:02.396 20:15:57 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@51 -- # : 0 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:17:02.396 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:17:02.396 20:15:57 json_config_extra_key -- nvmf/common.sh@55 -- # have_pci_nics=0 00:17:02.396 20:15:57 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:17:02.396 20:15:57 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:17:02.396 20:15:57 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:17:02.396 20:15:57 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:17:02.396 20:15:57 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:17:02.396 20:15:57 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1536') 00:17:02.396 20:15:57 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:17:02.396 20:15:57 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json') 00:17:02.396 INFO: launching applications... 00:17:02.396 20:15:57 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:17:02.396 20:15:57 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:17:02.396 20:15:57 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:17:02.396 20:15:57 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:17:02.396 20:15:57 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:17:02.396 20:15:57 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:17:02.396 20:15:57 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:17:02.396 20:15:57 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:17:02.396 20:15:57 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:17:02.396 20:15:57 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:17:02.396 20:15:57 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:17:02.396 20:15:57 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=57638 00:17:02.396 20:15:57 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:17:02.396 Waiting for target to run... 00:17:02.396 20:15:57 json_config_extra_key -- json_config/common.sh@21 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -s 1536 -r /var/tmp/spdk_tgt.sock --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:17:02.396 20:15:57 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 57638 /var/tmp/spdk_tgt.sock 00:17:02.396 20:15:57 json_config_extra_key -- common/autotest_common.sh@831 -- # '[' -z 57638 ']' 00:17:02.396 20:15:57 json_config_extra_key -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:17:02.396 20:15:57 json_config_extra_key -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:02.396 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:17:02.396 20:15:57 json_config_extra_key -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:17:02.396 20:15:57 json_config_extra_key -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:02.397 20:15:57 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:17:02.397 [2024-10-01 20:15:57.570631] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:17:02.397 [2024-10-01 20:15:57.571105] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1536 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57638 ] 00:17:02.990 [2024-10-01 20:15:58.183902] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:03.248 [2024-10-01 20:15:58.435769] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:04.184 00:17:04.184 INFO: shutting down applications... 00:17:04.184 20:15:59 json_config_extra_key -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:04.184 20:15:59 json_config_extra_key -- common/autotest_common.sh@864 -- # return 0 00:17:04.184 20:15:59 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:17:04.184 20:15:59 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:17:04.184 20:15:59 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:17:04.184 20:15:59 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:17:04.184 20:15:59 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:17:04.184 20:15:59 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 57638 ]] 00:17:04.184 20:15:59 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 57638 00:17:04.184 20:15:59 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:17:04.184 20:15:59 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:17:04.184 20:15:59 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 57638 00:17:04.184 20:15:59 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:17:04.751 20:15:59 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:17:04.751 20:15:59 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:17:04.751 20:15:59 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 57638 00:17:04.751 20:15:59 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:17:05.317 20:16:00 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:17:05.317 20:16:00 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:17:05.317 20:16:00 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 57638 00:17:05.317 20:16:00 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:17:05.883 20:16:00 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:17:05.883 20:16:00 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:17:05.883 20:16:00 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 57638 00:17:05.883 20:16:00 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:17:06.142 20:16:01 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:17:06.142 20:16:01 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:17:06.142 20:16:01 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 57638 00:17:06.142 20:16:01 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:17:06.709 20:16:01 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:17:06.709 20:16:01 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:17:06.709 20:16:01 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 57638 00:17:06.709 20:16:01 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:17:07.276 20:16:02 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:17:07.276 20:16:02 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:17:07.276 20:16:02 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 57638 00:17:07.276 20:16:02 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:17:07.844 20:16:02 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:17:07.844 20:16:02 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:17:07.844 20:16:02 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 57638 00:17:07.844 20:16:02 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:17:08.412 20:16:03 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:17:08.412 20:16:03 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:17:08.412 20:16:03 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 57638 00:17:08.412 SPDK target shutdown done 00:17:08.412 Success 00:17:08.412 20:16:03 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:17:08.412 20:16:03 json_config_extra_key -- json_config/common.sh@43 -- # break 00:17:08.412 20:16:03 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:17:08.412 20:16:03 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:17:08.412 20:16:03 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:17:08.412 00:17:08.412 real 0m6.154s 00:17:08.412 user 0m5.398s 00:17:08.412 sys 0m0.883s 00:17:08.412 20:16:03 json_config_extra_key -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:08.412 20:16:03 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:17:08.412 ************************************ 00:17:08.412 END TEST json_config_extra_key 00:17:08.412 ************************************ 00:17:08.412 20:16:03 -- spdk/autotest.sh@161 -- # run_test alias_rpc /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:17:08.412 20:16:03 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:17:08.412 20:16:03 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:08.412 20:16:03 -- common/autotest_common.sh@10 -- # set +x 00:17:08.412 ************************************ 00:17:08.412 START TEST alias_rpc 00:17:08.412 ************************************ 00:17:08.412 20:16:03 alias_rpc -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:17:08.412 * Looking for test storage... 00:17:08.412 * Found test storage at /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc 00:17:08.412 20:16:03 alias_rpc -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:17:08.412 20:16:03 alias_rpc -- common/autotest_common.sh@1681 -- # lcov --version 00:17:08.412 20:16:03 alias_rpc -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:17:08.412 20:16:03 alias_rpc -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@344 -- # case "$op" in 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@345 -- # : 1 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@365 -- # decimal 1 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@353 -- # local d=1 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@355 -- # echo 1 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@366 -- # decimal 2 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@353 -- # local d=2 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@355 -- # echo 2 00:17:08.412 20:16:03 alias_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:17:08.413 20:16:03 alias_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:08.413 20:16:03 alias_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:08.413 20:16:03 alias_rpc -- scripts/common.sh@368 -- # return 0 00:17:08.413 20:16:03 alias_rpc -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:08.413 20:16:03 alias_rpc -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:17:08.413 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:08.413 --rc genhtml_branch_coverage=1 00:17:08.413 --rc genhtml_function_coverage=1 00:17:08.413 --rc genhtml_legend=1 00:17:08.413 --rc geninfo_all_blocks=1 00:17:08.413 --rc geninfo_unexecuted_blocks=1 00:17:08.413 00:17:08.413 ' 00:17:08.413 20:16:03 alias_rpc -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:17:08.413 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:08.413 --rc genhtml_branch_coverage=1 00:17:08.413 --rc genhtml_function_coverage=1 00:17:08.413 --rc genhtml_legend=1 00:17:08.413 --rc geninfo_all_blocks=1 00:17:08.413 --rc geninfo_unexecuted_blocks=1 00:17:08.413 00:17:08.413 ' 00:17:08.413 20:16:03 alias_rpc -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:17:08.413 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:08.413 --rc genhtml_branch_coverage=1 00:17:08.413 --rc genhtml_function_coverage=1 00:17:08.413 --rc genhtml_legend=1 00:17:08.413 --rc geninfo_all_blocks=1 00:17:08.413 --rc geninfo_unexecuted_blocks=1 00:17:08.413 00:17:08.413 ' 00:17:08.413 20:16:03 alias_rpc -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:17:08.413 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:08.413 --rc genhtml_branch_coverage=1 00:17:08.413 --rc genhtml_function_coverage=1 00:17:08.413 --rc genhtml_legend=1 00:17:08.413 --rc geninfo_all_blocks=1 00:17:08.413 --rc geninfo_unexecuted_blocks=1 00:17:08.413 00:17:08.413 ' 00:17:08.413 20:16:03 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:17:08.413 20:16:03 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=57768 00:17:08.413 20:16:03 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 57768 00:17:08.413 20:16:03 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:17:08.413 20:16:03 alias_rpc -- common/autotest_common.sh@831 -- # '[' -z 57768 ']' 00:17:08.413 20:16:03 alias_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:08.413 20:16:03 alias_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:08.413 20:16:03 alias_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:08.413 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:08.413 20:16:03 alias_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:08.413 20:16:03 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:17:08.672 [2024-10-01 20:16:03.776423] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:17:08.672 [2024-10-01 20:16:03.777183] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57768 ] 00:17:08.931 [2024-10-01 20:16:03.957882] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:09.191 [2024-10-01 20:16:04.194501] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:10.127 20:16:05 alias_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:10.127 20:16:05 alias_rpc -- common/autotest_common.sh@864 -- # return 0 00:17:10.127 20:16:05 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py load_config -i 00:17:10.385 20:16:05 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 57768 00:17:10.385 20:16:05 alias_rpc -- common/autotest_common.sh@950 -- # '[' -z 57768 ']' 00:17:10.385 20:16:05 alias_rpc -- common/autotest_common.sh@954 -- # kill -0 57768 00:17:10.385 20:16:05 alias_rpc -- common/autotest_common.sh@955 -- # uname 00:17:10.385 20:16:05 alias_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:17:10.385 20:16:05 alias_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 57768 00:17:10.769 killing process with pid 57768 00:17:10.769 20:16:05 alias_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:17:10.769 20:16:05 alias_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:17:10.769 20:16:05 alias_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 57768' 00:17:10.769 20:16:05 alias_rpc -- common/autotest_common.sh@969 -- # kill 57768 00:17:10.769 20:16:05 alias_rpc -- common/autotest_common.sh@974 -- # wait 57768 00:17:14.058 ************************************ 00:17:14.058 END TEST alias_rpc 00:17:14.058 ************************************ 00:17:14.058 00:17:14.058 real 0m5.155s 00:17:14.058 user 0m5.214s 00:17:14.058 sys 0m0.731s 00:17:14.058 20:16:08 alias_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:14.058 20:16:08 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:17:14.058 20:16:08 -- spdk/autotest.sh@163 -- # [[ 0 -eq 0 ]] 00:17:14.058 20:16:08 -- spdk/autotest.sh@164 -- # run_test spdkcli_tcp /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:17:14.058 20:16:08 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:17:14.058 20:16:08 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:14.058 20:16:08 -- common/autotest_common.sh@10 -- # set +x 00:17:14.058 ************************************ 00:17:14.058 START TEST spdkcli_tcp 00:17:14.058 ************************************ 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:17:14.058 * Looking for test storage... 00:17:14.058 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@1681 -- # lcov --version 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@344 -- # case "$op" in 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@345 -- # : 1 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@365 -- # decimal 1 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@353 -- # local d=1 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@355 -- # echo 1 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@366 -- # decimal 2 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@353 -- # local d=2 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@355 -- # echo 2 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:14.058 20:16:08 spdkcli_tcp -- scripts/common.sh@368 -- # return 0 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:17:14.058 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:14.058 --rc genhtml_branch_coverage=1 00:17:14.058 --rc genhtml_function_coverage=1 00:17:14.058 --rc genhtml_legend=1 00:17:14.058 --rc geninfo_all_blocks=1 00:17:14.058 --rc geninfo_unexecuted_blocks=1 00:17:14.058 00:17:14.058 ' 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:17:14.058 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:14.058 --rc genhtml_branch_coverage=1 00:17:14.058 --rc genhtml_function_coverage=1 00:17:14.058 --rc genhtml_legend=1 00:17:14.058 --rc geninfo_all_blocks=1 00:17:14.058 --rc geninfo_unexecuted_blocks=1 00:17:14.058 00:17:14.058 ' 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:17:14.058 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:14.058 --rc genhtml_branch_coverage=1 00:17:14.058 --rc genhtml_function_coverage=1 00:17:14.058 --rc genhtml_legend=1 00:17:14.058 --rc geninfo_all_blocks=1 00:17:14.058 --rc geninfo_unexecuted_blocks=1 00:17:14.058 00:17:14.058 ' 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:17:14.058 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:14.058 --rc genhtml_branch_coverage=1 00:17:14.058 --rc genhtml_function_coverage=1 00:17:14.058 --rc genhtml_legend=1 00:17:14.058 --rc geninfo_all_blocks=1 00:17:14.058 --rc geninfo_unexecuted_blocks=1 00:17:14.058 00:17:14.058 ' 00:17:14.058 20:16:08 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:17:14.058 20:16:08 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:17:14.058 20:16:08 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:17:14.058 20:16:08 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:17:14.058 20:16:08 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:17:14.058 20:16:08 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:17:14.058 20:16:08 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:17:14.058 20:16:08 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=57886 00:17:14.058 20:16:08 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 57886 00:17:14.058 20:16:08 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@831 -- # '[' -z 57886 ']' 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:14.058 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:14.058 20:16:08 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:17:14.058 [2024-10-01 20:16:08.995218] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:17:14.058 [2024-10-01 20:16:08.995418] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid57886 ] 00:17:14.058 [2024-10-01 20:16:09.171588] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:14.317 [2024-10-01 20:16:09.415404] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:14.317 [2024-10-01 20:16:09.415432] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:17:15.306 20:16:10 spdkcli_tcp -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:15.306 20:16:10 spdkcli_tcp -- common/autotest_common.sh@864 -- # return 0 00:17:15.306 20:16:10 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=57914 00:17:15.306 20:16:10 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:17:15.306 20:16:10 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:17:15.887 [ 00:17:15.887 "bdev_malloc_delete", 00:17:15.887 "bdev_malloc_create", 00:17:15.887 "bdev_null_resize", 00:17:15.887 "bdev_null_delete", 00:17:15.887 "bdev_null_create", 00:17:15.887 "bdev_nvme_cuse_unregister", 00:17:15.887 "bdev_nvme_cuse_register", 00:17:15.887 "bdev_opal_new_user", 00:17:15.887 "bdev_opal_set_lock_state", 00:17:15.887 "bdev_opal_delete", 00:17:15.887 "bdev_opal_get_info", 00:17:15.887 "bdev_opal_create", 00:17:15.887 "bdev_nvme_opal_revert", 00:17:15.887 "bdev_nvme_opal_init", 00:17:15.887 "bdev_nvme_send_cmd", 00:17:15.887 "bdev_nvme_set_keys", 00:17:15.887 "bdev_nvme_get_path_iostat", 00:17:15.887 "bdev_nvme_get_mdns_discovery_info", 00:17:15.887 "bdev_nvme_stop_mdns_discovery", 00:17:15.887 "bdev_nvme_start_mdns_discovery", 00:17:15.887 "bdev_nvme_set_multipath_policy", 00:17:15.888 "bdev_nvme_set_preferred_path", 00:17:15.888 "bdev_nvme_get_io_paths", 00:17:15.888 "bdev_nvme_remove_error_injection", 00:17:15.888 "bdev_nvme_add_error_injection", 00:17:15.888 "bdev_nvme_get_discovery_info", 00:17:15.888 "bdev_nvme_stop_discovery", 00:17:15.888 "bdev_nvme_start_discovery", 00:17:15.888 "bdev_nvme_get_controller_health_info", 00:17:15.888 "bdev_nvme_disable_controller", 00:17:15.888 "bdev_nvme_enable_controller", 00:17:15.888 "bdev_nvme_reset_controller", 00:17:15.888 "bdev_nvme_get_transport_statistics", 00:17:15.888 "bdev_nvme_apply_firmware", 00:17:15.888 "bdev_nvme_detach_controller", 00:17:15.888 "bdev_nvme_get_controllers", 00:17:15.888 "bdev_nvme_attach_controller", 00:17:15.888 "bdev_nvme_set_hotplug", 00:17:15.888 "bdev_nvme_set_options", 00:17:15.888 "bdev_passthru_delete", 00:17:15.888 "bdev_passthru_create", 00:17:15.888 "bdev_lvol_set_parent_bdev", 00:17:15.888 "bdev_lvol_set_parent", 00:17:15.888 "bdev_lvol_check_shallow_copy", 00:17:15.888 "bdev_lvol_start_shallow_copy", 00:17:15.888 "bdev_lvol_grow_lvstore", 00:17:15.888 "bdev_lvol_get_lvols", 00:17:15.888 "bdev_lvol_get_lvstores", 00:17:15.888 "bdev_lvol_delete", 00:17:15.888 "bdev_lvol_set_read_only", 00:17:15.888 "bdev_lvol_resize", 00:17:15.888 "bdev_lvol_decouple_parent", 00:17:15.888 "bdev_lvol_inflate", 00:17:15.888 "bdev_lvol_rename", 00:17:15.888 "bdev_lvol_clone_bdev", 00:17:15.888 "bdev_lvol_clone", 00:17:15.888 "bdev_lvol_snapshot", 00:17:15.888 "bdev_lvol_create", 00:17:15.888 "bdev_lvol_delete_lvstore", 00:17:15.888 "bdev_lvol_rename_lvstore", 00:17:15.888 "bdev_lvol_create_lvstore", 00:17:15.888 "bdev_raid_set_options", 00:17:15.888 "bdev_raid_remove_base_bdev", 00:17:15.888 "bdev_raid_add_base_bdev", 00:17:15.888 "bdev_raid_delete", 00:17:15.888 "bdev_raid_create", 00:17:15.888 "bdev_raid_get_bdevs", 00:17:15.888 "bdev_error_inject_error", 00:17:15.888 "bdev_error_delete", 00:17:15.888 "bdev_error_create", 00:17:15.888 "bdev_split_delete", 00:17:15.888 "bdev_split_create", 00:17:15.888 "bdev_delay_delete", 00:17:15.888 "bdev_delay_create", 00:17:15.888 "bdev_delay_update_latency", 00:17:15.888 "bdev_zone_block_delete", 00:17:15.888 "bdev_zone_block_create", 00:17:15.888 "blobfs_create", 00:17:15.888 "blobfs_detect", 00:17:15.888 "blobfs_set_cache_size", 00:17:15.888 "bdev_aio_delete", 00:17:15.888 "bdev_aio_rescan", 00:17:15.888 "bdev_aio_create", 00:17:15.888 "bdev_ftl_set_property", 00:17:15.888 "bdev_ftl_get_properties", 00:17:15.888 "bdev_ftl_get_stats", 00:17:15.888 "bdev_ftl_unmap", 00:17:15.888 "bdev_ftl_unload", 00:17:15.888 "bdev_ftl_delete", 00:17:15.888 "bdev_ftl_load", 00:17:15.888 "bdev_ftl_create", 00:17:15.888 "bdev_virtio_attach_controller", 00:17:15.888 "bdev_virtio_scsi_get_devices", 00:17:15.888 "bdev_virtio_detach_controller", 00:17:15.888 "bdev_virtio_blk_set_hotplug", 00:17:15.888 "bdev_iscsi_delete", 00:17:15.888 "bdev_iscsi_create", 00:17:15.888 "bdev_iscsi_set_options", 00:17:15.888 "accel_error_inject_error", 00:17:15.888 "ioat_scan_accel_module", 00:17:15.888 "dsa_scan_accel_module", 00:17:15.888 "iaa_scan_accel_module", 00:17:15.888 "keyring_file_remove_key", 00:17:15.888 "keyring_file_add_key", 00:17:15.888 "keyring_linux_set_options", 00:17:15.888 "fsdev_aio_delete", 00:17:15.888 "fsdev_aio_create", 00:17:15.888 "iscsi_get_histogram", 00:17:15.888 "iscsi_enable_histogram", 00:17:15.888 "iscsi_set_options", 00:17:15.888 "iscsi_get_auth_groups", 00:17:15.888 "iscsi_auth_group_remove_secret", 00:17:15.888 "iscsi_auth_group_add_secret", 00:17:15.888 "iscsi_delete_auth_group", 00:17:15.888 "iscsi_create_auth_group", 00:17:15.888 "iscsi_set_discovery_auth", 00:17:15.888 "iscsi_get_options", 00:17:15.888 "iscsi_target_node_request_logout", 00:17:15.888 "iscsi_target_node_set_redirect", 00:17:15.888 "iscsi_target_node_set_auth", 00:17:15.888 "iscsi_target_node_add_lun", 00:17:15.888 "iscsi_get_stats", 00:17:15.888 "iscsi_get_connections", 00:17:15.888 "iscsi_portal_group_set_auth", 00:17:15.888 "iscsi_start_portal_group", 00:17:15.888 "iscsi_delete_portal_group", 00:17:15.888 "iscsi_create_portal_group", 00:17:15.888 "iscsi_get_portal_groups", 00:17:15.888 "iscsi_delete_target_node", 00:17:15.888 "iscsi_target_node_remove_pg_ig_maps", 00:17:15.888 "iscsi_target_node_add_pg_ig_maps", 00:17:15.888 "iscsi_create_target_node", 00:17:15.888 "iscsi_get_target_nodes", 00:17:15.888 "iscsi_delete_initiator_group", 00:17:15.888 "iscsi_initiator_group_remove_initiators", 00:17:15.888 "iscsi_initiator_group_add_initiators", 00:17:15.888 "iscsi_create_initiator_group", 00:17:15.888 "iscsi_get_initiator_groups", 00:17:15.888 "nvmf_set_crdt", 00:17:15.888 "nvmf_set_config", 00:17:15.888 "nvmf_set_max_subsystems", 00:17:15.888 "nvmf_stop_mdns_prr", 00:17:15.888 "nvmf_publish_mdns_prr", 00:17:15.888 "nvmf_subsystem_get_listeners", 00:17:15.888 "nvmf_subsystem_get_qpairs", 00:17:15.888 "nvmf_subsystem_get_controllers", 00:17:15.888 "nvmf_get_stats", 00:17:15.888 "nvmf_get_transports", 00:17:15.888 "nvmf_create_transport", 00:17:15.888 "nvmf_get_targets", 00:17:15.888 "nvmf_delete_target", 00:17:15.888 "nvmf_create_target", 00:17:15.888 "nvmf_subsystem_allow_any_host", 00:17:15.888 "nvmf_subsystem_set_keys", 00:17:15.888 "nvmf_subsystem_remove_host", 00:17:15.888 "nvmf_subsystem_add_host", 00:17:15.888 "nvmf_ns_remove_host", 00:17:15.888 "nvmf_ns_add_host", 00:17:15.888 "nvmf_subsystem_remove_ns", 00:17:15.888 "nvmf_subsystem_set_ns_ana_group", 00:17:15.888 "nvmf_subsystem_add_ns", 00:17:15.888 "nvmf_subsystem_listener_set_ana_state", 00:17:15.888 "nvmf_discovery_get_referrals", 00:17:15.888 "nvmf_discovery_remove_referral", 00:17:15.888 "nvmf_discovery_add_referral", 00:17:15.888 "nvmf_subsystem_remove_listener", 00:17:15.888 "nvmf_subsystem_add_listener", 00:17:15.888 "nvmf_delete_subsystem", 00:17:15.888 "nvmf_create_subsystem", 00:17:15.888 "nvmf_get_subsystems", 00:17:15.888 "env_dpdk_get_mem_stats", 00:17:15.888 "nbd_get_disks", 00:17:15.888 "nbd_stop_disk", 00:17:15.888 "nbd_start_disk", 00:17:15.888 "ublk_recover_disk", 00:17:15.888 "ublk_get_disks", 00:17:15.888 "ublk_stop_disk", 00:17:15.888 "ublk_start_disk", 00:17:15.888 "ublk_destroy_target", 00:17:15.888 "ublk_create_target", 00:17:15.888 "virtio_blk_create_transport", 00:17:15.888 "virtio_blk_get_transports", 00:17:15.888 "vhost_controller_set_coalescing", 00:17:15.888 "vhost_get_controllers", 00:17:15.888 "vhost_delete_controller", 00:17:15.888 "vhost_create_blk_controller", 00:17:15.888 "vhost_scsi_controller_remove_target", 00:17:15.888 "vhost_scsi_controller_add_target", 00:17:15.888 "vhost_start_scsi_controller", 00:17:15.888 "vhost_create_scsi_controller", 00:17:15.888 "thread_set_cpumask", 00:17:15.888 "scheduler_set_options", 00:17:15.888 "framework_get_governor", 00:17:15.888 "framework_get_scheduler", 00:17:15.888 "framework_set_scheduler", 00:17:15.888 "framework_get_reactors", 00:17:15.888 "thread_get_io_channels", 00:17:15.888 "thread_get_pollers", 00:17:15.888 "thread_get_stats", 00:17:15.888 "framework_monitor_context_switch", 00:17:15.888 "spdk_kill_instance", 00:17:15.888 "log_enable_timestamps", 00:17:15.888 "log_get_flags", 00:17:15.888 "log_clear_flag", 00:17:15.888 "log_set_flag", 00:17:15.888 "log_get_level", 00:17:15.888 "log_set_level", 00:17:15.888 "log_get_print_level", 00:17:15.888 "log_set_print_level", 00:17:15.888 "framework_enable_cpumask_locks", 00:17:15.888 "framework_disable_cpumask_locks", 00:17:15.888 "framework_wait_init", 00:17:15.888 "framework_start_init", 00:17:15.888 "scsi_get_devices", 00:17:15.888 "bdev_get_histogram", 00:17:15.888 "bdev_enable_histogram", 00:17:15.888 "bdev_set_qos_limit", 00:17:15.888 "bdev_set_qd_sampling_period", 00:17:15.888 "bdev_get_bdevs", 00:17:15.888 "bdev_reset_iostat", 00:17:15.888 "bdev_get_iostat", 00:17:15.888 "bdev_examine", 00:17:15.888 "bdev_wait_for_examine", 00:17:15.888 "bdev_set_options", 00:17:15.888 "accel_get_stats", 00:17:15.888 "accel_set_options", 00:17:15.888 "accel_set_driver", 00:17:15.888 "accel_crypto_key_destroy", 00:17:15.888 "accel_crypto_keys_get", 00:17:15.888 "accel_crypto_key_create", 00:17:15.888 "accel_assign_opc", 00:17:15.888 "accel_get_module_info", 00:17:15.888 "accel_get_opc_assignments", 00:17:15.888 "vmd_rescan", 00:17:15.888 "vmd_remove_device", 00:17:15.888 "vmd_enable", 00:17:15.888 "sock_get_default_impl", 00:17:15.888 "sock_set_default_impl", 00:17:15.888 "sock_impl_set_options", 00:17:15.888 "sock_impl_get_options", 00:17:15.888 "iobuf_get_stats", 00:17:15.888 "iobuf_set_options", 00:17:15.888 "keyring_get_keys", 00:17:15.888 "framework_get_pci_devices", 00:17:15.888 "framework_get_config", 00:17:15.888 "framework_get_subsystems", 00:17:15.888 "fsdev_set_opts", 00:17:15.888 "fsdev_get_opts", 00:17:15.888 "trace_get_info", 00:17:15.888 "trace_get_tpoint_group_mask", 00:17:15.888 "trace_disable_tpoint_group", 00:17:15.888 "trace_enable_tpoint_group", 00:17:15.889 "trace_clear_tpoint_mask", 00:17:15.889 "trace_set_tpoint_mask", 00:17:15.889 "notify_get_notifications", 00:17:15.889 "notify_get_types", 00:17:15.889 "spdk_get_version", 00:17:15.889 "rpc_get_methods" 00:17:15.889 ] 00:17:15.889 20:16:10 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:17:15.889 20:16:10 spdkcli_tcp -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:15.889 20:16:10 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:17:15.889 20:16:10 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:17:15.889 20:16:10 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 57886 00:17:15.889 20:16:10 spdkcli_tcp -- common/autotest_common.sh@950 -- # '[' -z 57886 ']' 00:17:15.889 20:16:10 spdkcli_tcp -- common/autotest_common.sh@954 -- # kill -0 57886 00:17:15.889 20:16:10 spdkcli_tcp -- common/autotest_common.sh@955 -- # uname 00:17:15.889 20:16:10 spdkcli_tcp -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:17:15.889 20:16:10 spdkcli_tcp -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 57886 00:17:15.889 killing process with pid 57886 00:17:15.889 20:16:10 spdkcli_tcp -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:17:15.889 20:16:10 spdkcli_tcp -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:17:15.889 20:16:10 spdkcli_tcp -- common/autotest_common.sh@968 -- # echo 'killing process with pid 57886' 00:17:15.889 20:16:10 spdkcli_tcp -- common/autotest_common.sh@969 -- # kill 57886 00:17:15.889 20:16:10 spdkcli_tcp -- common/autotest_common.sh@974 -- # wait 57886 00:17:19.175 ************************************ 00:17:19.175 END TEST spdkcli_tcp 00:17:19.175 ************************************ 00:17:19.175 00:17:19.175 real 0m5.244s 00:17:19.175 user 0m9.349s 00:17:19.175 sys 0m0.766s 00:17:19.175 20:16:13 spdkcli_tcp -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:19.175 20:16:13 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:17:19.175 20:16:13 -- spdk/autotest.sh@167 -- # run_test dpdk_mem_utility /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:17:19.175 20:16:13 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:17:19.175 20:16:13 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:19.175 20:16:13 -- common/autotest_common.sh@10 -- # set +x 00:17:19.175 ************************************ 00:17:19.175 START TEST dpdk_mem_utility 00:17:19.175 ************************************ 00:17:19.175 20:16:13 dpdk_mem_utility -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:17:19.175 * Looking for test storage... 00:17:19.175 * Found test storage at /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility 00:17:19.175 20:16:14 dpdk_mem_utility -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:17:19.175 20:16:14 dpdk_mem_utility -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:17:19.175 20:16:14 dpdk_mem_utility -- common/autotest_common.sh@1681 -- # lcov --version 00:17:19.175 20:16:14 dpdk_mem_utility -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:17:19.175 20:16:14 dpdk_mem_utility -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:19.175 20:16:14 dpdk_mem_utility -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:19.175 20:16:14 dpdk_mem_utility -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:19.175 20:16:14 dpdk_mem_utility -- scripts/common.sh@336 -- # IFS=.-: 00:17:19.175 20:16:14 dpdk_mem_utility -- scripts/common.sh@336 -- # read -ra ver1 00:17:19.175 20:16:14 dpdk_mem_utility -- scripts/common.sh@337 -- # IFS=.-: 00:17:19.175 20:16:14 dpdk_mem_utility -- scripts/common.sh@337 -- # read -ra ver2 00:17:19.175 20:16:14 dpdk_mem_utility -- scripts/common.sh@338 -- # local 'op=<' 00:17:19.175 20:16:14 dpdk_mem_utility -- scripts/common.sh@340 -- # ver1_l=2 00:17:19.175 20:16:14 dpdk_mem_utility -- scripts/common.sh@341 -- # ver2_l=1 00:17:19.175 20:16:14 dpdk_mem_utility -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:19.175 20:16:14 dpdk_mem_utility -- scripts/common.sh@344 -- # case "$op" in 00:17:19.175 20:16:14 dpdk_mem_utility -- scripts/common.sh@345 -- # : 1 00:17:19.175 20:16:14 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:19.175 20:16:14 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:19.176 20:16:14 dpdk_mem_utility -- scripts/common.sh@365 -- # decimal 1 00:17:19.176 20:16:14 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=1 00:17:19.176 20:16:14 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:19.176 20:16:14 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 1 00:17:19.176 20:16:14 dpdk_mem_utility -- scripts/common.sh@365 -- # ver1[v]=1 00:17:19.176 20:16:14 dpdk_mem_utility -- scripts/common.sh@366 -- # decimal 2 00:17:19.176 20:16:14 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=2 00:17:19.176 20:16:14 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:19.176 20:16:14 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 2 00:17:19.176 20:16:14 dpdk_mem_utility -- scripts/common.sh@366 -- # ver2[v]=2 00:17:19.176 20:16:14 dpdk_mem_utility -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:19.176 20:16:14 dpdk_mem_utility -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:19.176 20:16:14 dpdk_mem_utility -- scripts/common.sh@368 -- # return 0 00:17:19.176 20:16:14 dpdk_mem_utility -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:19.176 20:16:14 dpdk_mem_utility -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:17:19.176 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:19.176 --rc genhtml_branch_coverage=1 00:17:19.176 --rc genhtml_function_coverage=1 00:17:19.176 --rc genhtml_legend=1 00:17:19.176 --rc geninfo_all_blocks=1 00:17:19.176 --rc geninfo_unexecuted_blocks=1 00:17:19.176 00:17:19.176 ' 00:17:19.176 20:16:14 dpdk_mem_utility -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:17:19.176 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:19.176 --rc genhtml_branch_coverage=1 00:17:19.176 --rc genhtml_function_coverage=1 00:17:19.176 --rc genhtml_legend=1 00:17:19.176 --rc geninfo_all_blocks=1 00:17:19.176 --rc geninfo_unexecuted_blocks=1 00:17:19.176 00:17:19.176 ' 00:17:19.176 20:16:14 dpdk_mem_utility -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:17:19.176 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:19.176 --rc genhtml_branch_coverage=1 00:17:19.176 --rc genhtml_function_coverage=1 00:17:19.176 --rc genhtml_legend=1 00:17:19.176 --rc geninfo_all_blocks=1 00:17:19.176 --rc geninfo_unexecuted_blocks=1 00:17:19.176 00:17:19.176 ' 00:17:19.176 20:16:14 dpdk_mem_utility -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:17:19.176 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:19.176 --rc genhtml_branch_coverage=1 00:17:19.176 --rc genhtml_function_coverage=1 00:17:19.176 --rc genhtml_legend=1 00:17:19.176 --rc geninfo_all_blocks=1 00:17:19.176 --rc geninfo_unexecuted_blocks=1 00:17:19.176 00:17:19.176 ' 00:17:19.176 20:16:14 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:17:19.176 20:16:14 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=58019 00:17:19.176 20:16:14 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:17:19.176 20:16:14 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 58019 00:17:19.176 20:16:14 dpdk_mem_utility -- common/autotest_common.sh@831 -- # '[' -z 58019 ']' 00:17:19.176 20:16:14 dpdk_mem_utility -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:19.176 20:16:14 dpdk_mem_utility -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:19.176 20:16:14 dpdk_mem_utility -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:19.176 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:19.176 20:16:14 dpdk_mem_utility -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:19.176 20:16:14 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:17:19.176 [2024-10-01 20:16:14.273474] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:17:19.176 [2024-10-01 20:16:14.274068] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58019 ] 00:17:19.434 [2024-10-01 20:16:14.441408] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:19.434 [2024-10-01 20:16:14.682634] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:20.814 20:16:15 dpdk_mem_utility -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:20.814 20:16:15 dpdk_mem_utility -- common/autotest_common.sh@864 -- # return 0 00:17:20.814 20:16:15 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:17:20.814 20:16:15 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:17:20.814 20:16:15 dpdk_mem_utility -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:20.814 20:16:15 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:17:20.814 { 00:17:20.814 "filename": "/tmp/spdk_mem_dump.txt" 00:17:20.814 } 00:17:20.814 20:16:15 dpdk_mem_utility -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:20.814 20:16:15 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:17:20.814 DPDK memory size 1106.000000 MiB in 1 heap(s) 00:17:20.814 1 heaps totaling size 1106.000000 MiB 00:17:20.814 size: 1106.000000 MiB heap id: 0 00:17:20.814 end heaps---------- 00:17:20.814 9 mempools totaling size 883.273621 MiB 00:17:20.814 size: 333.169250 MiB name: bdev_io_58019 00:17:20.814 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:17:20.814 size: 158.602051 MiB name: PDU_data_out_Pool 00:17:20.814 size: 51.011292 MiB name: evtpool_58019 00:17:20.814 size: 50.003479 MiB name: msgpool_58019 00:17:20.814 size: 36.509338 MiB name: fsdev_io_58019 00:17:20.815 size: 21.763794 MiB name: PDU_Pool 00:17:20.815 size: 19.513306 MiB name: SCSI_TASK_Pool 00:17:20.815 size: 0.026123 MiB name: Session_Pool 00:17:20.815 end mempools------- 00:17:20.815 6 memzones totaling size 4.142822 MiB 00:17:20.815 size: 1.000366 MiB name: RG_ring_0_58019 00:17:20.815 size: 1.000366 MiB name: RG_ring_1_58019 00:17:20.815 size: 1.000366 MiB name: RG_ring_4_58019 00:17:20.815 size: 1.000366 MiB name: RG_ring_5_58019 00:17:20.815 size: 0.125366 MiB name: RG_ring_2_58019 00:17:20.815 size: 0.015991 MiB name: RG_ring_3_58019 00:17:20.815 end memzones------- 00:17:20.815 20:16:15 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py -m 0 00:17:20.815 heap id: 0 total size: 1106.000000 MiB number of busy elements: 370 number of free elements: 19 00:17:20.815 list of free elements. size: 19.276367 MiB 00:17:20.815 element at address: 0x200000400000 with size: 1.999451 MiB 00:17:20.815 element at address: 0x200000800000 with size: 1.996887 MiB 00:17:20.815 element at address: 0x200009600000 with size: 1.995972 MiB 00:17:20.815 element at address: 0x20000d800000 with size: 1.995972 MiB 00:17:20.815 element at address: 0x200007000000 with size: 1.991028 MiB 00:17:20.815 element at address: 0x20002af00040 with size: 0.999939 MiB 00:17:20.815 element at address: 0x20002b300040 with size: 0.999939 MiB 00:17:20.815 element at address: 0x20002b400000 with size: 0.999084 MiB 00:17:20.815 element at address: 0x200044000000 with size: 0.994324 MiB 00:17:20.815 element at address: 0x20002b700040 with size: 0.936401 MiB 00:17:20.815 element at address: 0x200000200000 with size: 0.829224 MiB 00:17:20.815 element at address: 0x20002ce00000 with size: 0.563904 MiB 00:17:20.815 element at address: 0x20002b000000 with size: 0.489197 MiB 00:17:20.815 element at address: 0x20002b800000 with size: 0.485413 MiB 00:17:20.815 element at address: 0x200003e00000 with size: 0.479431 MiB 00:17:20.815 element at address: 0x20002ac00000 with size: 0.456421 MiB 00:17:20.815 element at address: 0x20003a200000 with size: 0.390442 MiB 00:17:20.815 element at address: 0x200003a00000 with size: 0.350647 MiB 00:17:20.815 element at address: 0x200015e00000 with size: 0.322693 MiB 00:17:20.815 list of standard malloc elements. size: 199.301147 MiB 00:17:20.815 element at address: 0x20000d9fef80 with size: 132.000183 MiB 00:17:20.815 element at address: 0x2000097fef80 with size: 64.000183 MiB 00:17:20.815 element at address: 0x20002adfff80 with size: 1.000183 MiB 00:17:20.815 element at address: 0x20002b1fff80 with size: 1.000183 MiB 00:17:20.815 element at address: 0x20002b5fff80 with size: 1.000183 MiB 00:17:20.815 element at address: 0x2000003d9e80 with size: 0.140808 MiB 00:17:20.815 element at address: 0x20002b7eff40 with size: 0.062683 MiB 00:17:20.815 element at address: 0x2000003fdf40 with size: 0.007996 MiB 00:17:20.815 element at address: 0x20000d7ff040 with size: 0.000427 MiB 00:17:20.815 element at address: 0x20002b7efdc0 with size: 0.000366 MiB 00:17:20.815 element at address: 0x200015dff040 with size: 0.000305 MiB 00:17:20.815 element at address: 0x2000002d4480 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d4580 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d4680 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d4780 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d4880 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d4980 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d4a80 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d4b80 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d4c80 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d4d80 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d4e80 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d4f80 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d5080 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d5180 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d5280 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d5380 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d5480 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d5580 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d5680 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d5780 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d5880 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d5980 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d5a80 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d5b80 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d5c80 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d5d80 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d5e80 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d6100 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d6200 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d6300 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d6400 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d6500 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d6600 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d6700 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d6800 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d6900 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d6a00 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d6b00 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d6c00 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d6d00 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d6e00 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d6f00 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d7000 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d7100 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d7200 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d7300 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d7400 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d7500 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d7600 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d7700 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d7800 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d7900 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d7a00 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000002d7b00 with size: 0.000244 MiB 00:17:20.815 element at address: 0x2000003d9d80 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7e0c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7e1c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7e2c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7e3c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7e4c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7e5c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7e6c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7e7c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7e8c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7e9c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7eac0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7ebc0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7ecc0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7edc0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7eec0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7efc0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7f0c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7f1c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7f2c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003a7f3c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003aff700 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003aff980 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003affa80 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7abc0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7acc0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7adc0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7aec0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7afc0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7b0c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7b1c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7b2c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7b3c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7b4c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7b5c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7b6c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7b7c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7b8c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7b9c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7bac0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7bbc0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7bcc0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7bdc0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7bec0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7bfc0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7c0c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7c1c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7c2c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7c3c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7c4c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7c5c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7c6c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7c7c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7c8c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7c9c0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7cac0 with size: 0.000244 MiB 00:17:20.815 element at address: 0x200003e7cbc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7ccc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7cdc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7cec0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7cfc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7d0c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7d1c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7d2c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7d3c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7d4c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7d5c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7d6c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7d7c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7d8c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7d9c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7dac0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7dbc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7dcc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7ddc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7dec0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7dfc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7e0c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7e1c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7e2c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7e3c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7e4c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7e5c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7e6c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7e7c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7e8c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7e9c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7eac0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7ebc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003e7ecc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200003eff000 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20000d7ff200 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20000d7ff300 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20000d7ff400 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20000d7ff500 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20000d7ff600 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20000d7ff700 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20000d7ff800 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20000d7ff900 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20000d7ffa00 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20000d7ffb00 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20000d7ffc00 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20000d7ffd00 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20000d7ffe00 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20000d7fff00 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200015dff180 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200015dff280 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200015dff380 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200015dff480 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200015dff580 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200015dff680 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200015dff780 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200015dff880 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200015dff980 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200015dffa80 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200015dffb80 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200015dffc80 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200015dfff00 with size: 0.000244 MiB 00:17:20.816 element at address: 0x200015e529c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ac74d80 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ac74e80 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ac74f80 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ac75080 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ac75180 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ac75280 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ac75380 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ac75480 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ac75580 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ac75680 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ac75780 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002acfdd00 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002b07d3c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002b07d4c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002b07d5c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002b07d6c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002b07d7c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002b07d8c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002b07d9c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002b0fdd00 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002b4ffc40 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002b7efbc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002b7efcc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002b8bc680 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce905c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce906c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce907c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce908c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce909c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce90ac0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce90bc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce90cc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce90dc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce90ec0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce90fc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce910c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce911c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce912c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce913c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce914c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce915c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce916c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce917c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce918c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce919c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce91ac0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce91bc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce91cc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce91dc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce91ec0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce91fc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce920c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce921c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce922c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce923c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce924c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce925c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce926c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce927c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce928c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce929c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce92ac0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce92bc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce92cc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce92dc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce92ec0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce92fc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce930c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce931c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce932c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce933c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce934c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce935c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce936c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce937c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce938c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce939c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce93ac0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce93bc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce93cc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce93dc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce93ec0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce93fc0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce940c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce941c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce942c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce943c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce944c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce945c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce946c0 with size: 0.000244 MiB 00:17:20.816 element at address: 0x20002ce947c0 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20002ce948c0 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20002ce949c0 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20002ce94ac0 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20002ce94bc0 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20002ce94cc0 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20002ce94dc0 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20002ce94ec0 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20002ce94fc0 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20002ce950c0 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20002ce951c0 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20002ce952c0 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20002ce953c0 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a263f40 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a264040 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26ad00 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26af80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26b080 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26b180 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26b280 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26b380 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26b480 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26b580 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26b680 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26b780 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26b880 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26b980 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26ba80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26bb80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26bc80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26bd80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26be80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26bf80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26c080 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26c180 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26c280 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26c380 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26c480 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26c580 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26c680 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26c780 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26c880 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26c980 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26ca80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26cb80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26cc80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26cd80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26ce80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26cf80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26d080 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26d180 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26d280 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26d380 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26d480 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26d580 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26d680 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26d780 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26d880 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26d980 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26da80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26db80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26dc80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26dd80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26de80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26df80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26e080 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26e180 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26e280 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26e380 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26e480 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26e580 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26e680 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26e780 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26e880 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26e980 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26ea80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26eb80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26ec80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26ed80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26ee80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26ef80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26f080 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26f180 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26f280 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26f380 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26f480 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26f580 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26f680 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26f780 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26f880 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26f980 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26fa80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26fb80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26fc80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26fd80 with size: 0.000244 MiB 00:17:20.817 element at address: 0x20003a26fe80 with size: 0.000244 MiB 00:17:20.817 list of memzone associated elements. size: 887.422485 MiB 00:17:20.817 element at address: 0x200015f54c40 with size: 332.668884 MiB 00:17:20.817 associated memzone info: size: 332.668701 MiB name: MP_bdev_io_58019_0 00:17:20.817 element at address: 0x20002ce954c0 with size: 211.416809 MiB 00:17:20.817 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:17:20.817 element at address: 0x20003a26ff80 with size: 157.562622 MiB 00:17:20.817 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:17:20.817 element at address: 0x2000009ff340 with size: 48.003113 MiB 00:17:20.817 associated memzone info: size: 48.002930 MiB name: MP_evtpool_58019_0 00:17:20.817 element at address: 0x200003fff340 with size: 48.003113 MiB 00:17:20.817 associated memzone info: size: 48.002930 MiB name: MP_msgpool_58019_0 00:17:20.817 element at address: 0x2000071fdb40 with size: 36.008972 MiB 00:17:20.817 associated memzone info: size: 36.008789 MiB name: MP_fsdev_io_58019_0 00:17:20.817 element at address: 0x20002b9be900 with size: 20.255615 MiB 00:17:20.817 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:17:20.817 element at address: 0x2000441feb00 with size: 18.005127 MiB 00:17:20.817 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:17:20.817 element at address: 0x2000005ffdc0 with size: 2.000549 MiB 00:17:20.817 associated memzone info: size: 2.000366 MiB name: RG_MP_evtpool_58019 00:17:20.817 element at address: 0x200003bffdc0 with size: 2.000549 MiB 00:17:20.817 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_58019 00:17:20.817 element at address: 0x2000002d7c00 with size: 1.008179 MiB 00:17:20.817 associated memzone info: size: 1.007996 MiB name: MP_evtpool_58019 00:17:20.817 element at address: 0x20002b0fde00 with size: 1.008179 MiB 00:17:20.817 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:17:20.817 element at address: 0x20002b8bc780 with size: 1.008179 MiB 00:17:20.817 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:17:20.817 element at address: 0x20002acfde00 with size: 1.008179 MiB 00:17:20.817 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:17:20.817 element at address: 0x200015e52ac0 with size: 1.008179 MiB 00:17:20.817 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:17:20.817 element at address: 0x200003eff100 with size: 1.000549 MiB 00:17:20.817 associated memzone info: size: 1.000366 MiB name: RG_ring_0_58019 00:17:20.817 element at address: 0x200003affb80 with size: 1.000549 MiB 00:17:20.817 associated memzone info: size: 1.000366 MiB name: RG_ring_1_58019 00:17:20.817 element at address: 0x20002b4ffd40 with size: 1.000549 MiB 00:17:20.817 associated memzone info: size: 1.000366 MiB name: RG_ring_4_58019 00:17:20.817 element at address: 0x2000440fe8c0 with size: 1.000549 MiB 00:17:20.817 associated memzone info: size: 1.000366 MiB name: RG_ring_5_58019 00:17:20.817 element at address: 0x200003a7f4c0 with size: 0.500549 MiB 00:17:20.817 associated memzone info: size: 0.500366 MiB name: RG_MP_fsdev_io_58019 00:17:20.817 element at address: 0x200003e7edc0 with size: 0.500549 MiB 00:17:20.817 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_58019 00:17:20.817 element at address: 0x20002b07dac0 with size: 0.500549 MiB 00:17:20.817 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:17:20.817 element at address: 0x20002ac75880 with size: 0.500549 MiB 00:17:20.817 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:17:20.817 element at address: 0x20002b87c440 with size: 0.250549 MiB 00:17:20.817 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:17:20.817 element at address: 0x200003a5de80 with size: 0.125549 MiB 00:17:20.817 associated memzone info: size: 0.125366 MiB name: RG_ring_2_58019 00:17:20.817 element at address: 0x20002acf5ac0 with size: 0.031799 MiB 00:17:20.817 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:17:20.817 element at address: 0x20003a264140 with size: 0.023804 MiB 00:17:20.817 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:17:20.817 element at address: 0x200003a59c40 with size: 0.016174 MiB 00:17:20.817 associated memzone info: size: 0.015991 MiB name: RG_ring_3_58019 00:17:20.817 element at address: 0x20003a26a2c0 with size: 0.002502 MiB 00:17:20.818 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:17:20.818 element at address: 0x2000002d5f80 with size: 0.000366 MiB 00:17:20.818 associated memzone info: size: 0.000183 MiB name: MP_msgpool_58019 00:17:20.818 element at address: 0x200003aff800 with size: 0.000366 MiB 00:17:20.818 associated memzone info: size: 0.000183 MiB name: MP_fsdev_io_58019 00:17:20.818 element at address: 0x200015dffd80 with size: 0.000366 MiB 00:17:20.818 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_58019 00:17:20.818 element at address: 0x20003a26ae00 with size: 0.000366 MiB 00:17:20.818 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:17:20.818 20:16:15 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:17:20.818 20:16:15 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 58019 00:17:20.818 20:16:15 dpdk_mem_utility -- common/autotest_common.sh@950 -- # '[' -z 58019 ']' 00:17:20.818 20:16:15 dpdk_mem_utility -- common/autotest_common.sh@954 -- # kill -0 58019 00:17:20.818 20:16:15 dpdk_mem_utility -- common/autotest_common.sh@955 -- # uname 00:17:20.818 20:16:15 dpdk_mem_utility -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:17:20.818 20:16:15 dpdk_mem_utility -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 58019 00:17:20.818 killing process with pid 58019 00:17:20.818 20:16:16 dpdk_mem_utility -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:17:20.818 20:16:16 dpdk_mem_utility -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:17:20.818 20:16:16 dpdk_mem_utility -- common/autotest_common.sh@968 -- # echo 'killing process with pid 58019' 00:17:20.818 20:16:16 dpdk_mem_utility -- common/autotest_common.sh@969 -- # kill 58019 00:17:20.818 20:16:16 dpdk_mem_utility -- common/autotest_common.sh@974 -- # wait 58019 00:17:24.098 00:17:24.098 real 0m5.009s 00:17:24.098 user 0m4.852s 00:17:24.098 sys 0m0.750s 00:17:24.098 20:16:18 dpdk_mem_utility -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:24.098 ************************************ 00:17:24.098 END TEST dpdk_mem_utility 00:17:24.098 ************************************ 00:17:24.098 20:16:18 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:17:24.098 20:16:19 -- spdk/autotest.sh@168 -- # run_test event /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:17:24.098 20:16:19 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:17:24.098 20:16:19 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:24.098 20:16:19 -- common/autotest_common.sh@10 -- # set +x 00:17:24.098 ************************************ 00:17:24.098 START TEST event 00:17:24.098 ************************************ 00:17:24.098 20:16:19 event -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:17:24.098 * Looking for test storage... 00:17:24.098 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:17:24.098 20:16:19 event -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:17:24.098 20:16:19 event -- common/autotest_common.sh@1681 -- # lcov --version 00:17:24.098 20:16:19 event -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:17:24.098 20:16:19 event -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:17:24.098 20:16:19 event -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:24.098 20:16:19 event -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:24.098 20:16:19 event -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:24.098 20:16:19 event -- scripts/common.sh@336 -- # IFS=.-: 00:17:24.098 20:16:19 event -- scripts/common.sh@336 -- # read -ra ver1 00:17:24.098 20:16:19 event -- scripts/common.sh@337 -- # IFS=.-: 00:17:24.098 20:16:19 event -- scripts/common.sh@337 -- # read -ra ver2 00:17:24.098 20:16:19 event -- scripts/common.sh@338 -- # local 'op=<' 00:17:24.098 20:16:19 event -- scripts/common.sh@340 -- # ver1_l=2 00:17:24.098 20:16:19 event -- scripts/common.sh@341 -- # ver2_l=1 00:17:24.098 20:16:19 event -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:24.098 20:16:19 event -- scripts/common.sh@344 -- # case "$op" in 00:17:24.098 20:16:19 event -- scripts/common.sh@345 -- # : 1 00:17:24.098 20:16:19 event -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:24.099 20:16:19 event -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:24.099 20:16:19 event -- scripts/common.sh@365 -- # decimal 1 00:17:24.099 20:16:19 event -- scripts/common.sh@353 -- # local d=1 00:17:24.099 20:16:19 event -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:24.099 20:16:19 event -- scripts/common.sh@355 -- # echo 1 00:17:24.099 20:16:19 event -- scripts/common.sh@365 -- # ver1[v]=1 00:17:24.099 20:16:19 event -- scripts/common.sh@366 -- # decimal 2 00:17:24.099 20:16:19 event -- scripts/common.sh@353 -- # local d=2 00:17:24.099 20:16:19 event -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:24.099 20:16:19 event -- scripts/common.sh@355 -- # echo 2 00:17:24.099 20:16:19 event -- scripts/common.sh@366 -- # ver2[v]=2 00:17:24.099 20:16:19 event -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:24.099 20:16:19 event -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:24.099 20:16:19 event -- scripts/common.sh@368 -- # return 0 00:17:24.099 20:16:19 event -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:24.099 20:16:19 event -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:17:24.099 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:24.099 --rc genhtml_branch_coverage=1 00:17:24.099 --rc genhtml_function_coverage=1 00:17:24.099 --rc genhtml_legend=1 00:17:24.099 --rc geninfo_all_blocks=1 00:17:24.099 --rc geninfo_unexecuted_blocks=1 00:17:24.099 00:17:24.099 ' 00:17:24.099 20:16:19 event -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:17:24.099 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:24.099 --rc genhtml_branch_coverage=1 00:17:24.099 --rc genhtml_function_coverage=1 00:17:24.099 --rc genhtml_legend=1 00:17:24.099 --rc geninfo_all_blocks=1 00:17:24.099 --rc geninfo_unexecuted_blocks=1 00:17:24.099 00:17:24.099 ' 00:17:24.099 20:16:19 event -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:17:24.099 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:24.099 --rc genhtml_branch_coverage=1 00:17:24.099 --rc genhtml_function_coverage=1 00:17:24.099 --rc genhtml_legend=1 00:17:24.099 --rc geninfo_all_blocks=1 00:17:24.099 --rc geninfo_unexecuted_blocks=1 00:17:24.099 00:17:24.099 ' 00:17:24.099 20:16:19 event -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:17:24.099 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:24.099 --rc genhtml_branch_coverage=1 00:17:24.099 --rc genhtml_function_coverage=1 00:17:24.099 --rc genhtml_legend=1 00:17:24.099 --rc geninfo_all_blocks=1 00:17:24.099 --rc geninfo_unexecuted_blocks=1 00:17:24.099 00:17:24.099 ' 00:17:24.099 20:16:19 event -- event/event.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:17:24.099 20:16:19 event -- bdev/nbd_common.sh@6 -- # set -e 00:17:24.099 20:16:19 event -- event/event.sh@45 -- # run_test event_perf /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:17:24.099 20:16:19 event -- common/autotest_common.sh@1101 -- # '[' 6 -le 1 ']' 00:17:24.099 20:16:19 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:24.099 20:16:19 event -- common/autotest_common.sh@10 -- # set +x 00:17:24.099 ************************************ 00:17:24.099 START TEST event_perf 00:17:24.099 ************************************ 00:17:24.099 20:16:19 event.event_perf -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:17:24.099 Running I/O for 1 seconds...[2024-10-01 20:16:19.279521] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:17:24.099 [2024-10-01 20:16:19.279924] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58138 ] 00:17:24.357 [2024-10-01 20:16:19.462808] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 4 00:17:24.615 [2024-10-01 20:16:19.747642] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:17:24.615 [2024-10-01 20:16:19.747801] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:17:24.615 [2024-10-01 20:16:19.748171] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:24.615 [2024-10-01 20:16:19.748176] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 3 00:17:25.988 Running I/O for 1 seconds... 00:17:25.988 lcore 0: 185780 00:17:25.988 lcore 1: 185781 00:17:25.988 lcore 2: 185775 00:17:25.988 lcore 3: 185778 00:17:25.988 done. 00:17:25.988 00:17:25.988 real 0m1.906s 00:17:25.988 user 0m4.637s 00:17:25.988 sys 0m0.140s 00:17:25.988 ************************************ 00:17:25.988 END TEST event_perf 00:17:25.988 ************************************ 00:17:25.988 20:16:21 event.event_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:25.988 20:16:21 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:17:25.988 20:16:21 event -- event/event.sh@46 -- # run_test event_reactor /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:17:25.988 20:16:21 event -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:17:25.988 20:16:21 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:25.988 20:16:21 event -- common/autotest_common.sh@10 -- # set +x 00:17:25.988 ************************************ 00:17:25.988 START TEST event_reactor 00:17:25.988 ************************************ 00:17:25.988 20:16:21 event.event_reactor -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:17:25.988 [2024-10-01 20:16:21.234542] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:17:25.988 [2024-10-01 20:16:21.234731] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58183 ] 00:17:26.247 [2024-10-01 20:16:21.399109] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:26.505 [2024-10-01 20:16:21.642451] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:27.880 test_start 00:17:27.880 oneshot 00:17:27.881 tick 100 00:17:27.881 tick 100 00:17:27.881 tick 250 00:17:27.881 tick 100 00:17:27.881 tick 100 00:17:27.881 tick 100 00:17:27.881 tick 250 00:17:27.881 tick 500 00:17:27.881 tick 100 00:17:27.881 tick 100 00:17:27.881 tick 250 00:17:27.881 tick 100 00:17:27.881 tick 100 00:17:27.881 test_end 00:17:27.881 00:17:27.881 real 0m1.848s 00:17:27.881 user 0m1.625s 00:17:27.881 sys 0m0.113s 00:17:27.881 ************************************ 00:17:27.881 END TEST event_reactor 00:17:27.881 ************************************ 00:17:27.881 20:16:23 event.event_reactor -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:27.881 20:16:23 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:17:27.881 20:16:23 event -- event/event.sh@47 -- # run_test event_reactor_perf /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:17:27.881 20:16:23 event -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:17:27.881 20:16:23 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:27.881 20:16:23 event -- common/autotest_common.sh@10 -- # set +x 00:17:27.881 ************************************ 00:17:27.881 START TEST event_reactor_perf 00:17:27.881 ************************************ 00:17:27.881 20:16:23 event.event_reactor_perf -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:17:28.140 [2024-10-01 20:16:23.143582] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:17:28.140 [2024-10-01 20:16:23.143814] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58225 ] 00:17:28.140 [2024-10-01 20:16:23.322032] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:28.399 [2024-10-01 20:16:23.559415] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:29.778 test_start 00:17:29.778 test_end 00:17:29.778 Performance: 283842 events per second 00:17:29.778 00:17:29.778 real 0m1.870s 00:17:29.778 user 0m1.643s 00:17:29.778 sys 0m0.116s 00:17:29.778 20:16:24 event.event_reactor_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:29.778 20:16:24 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:17:29.778 ************************************ 00:17:29.778 END TEST event_reactor_perf 00:17:29.778 ************************************ 00:17:29.778 20:16:25 event -- event/event.sh@49 -- # uname -s 00:17:29.779 20:16:25 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:17:29.779 20:16:25 event -- event/event.sh@50 -- # run_test event_scheduler /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:17:29.779 20:16:25 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:17:29.779 20:16:25 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:29.779 20:16:25 event -- common/autotest_common.sh@10 -- # set +x 00:17:29.779 ************************************ 00:17:29.779 START TEST event_scheduler 00:17:29.779 ************************************ 00:17:29.779 20:16:25 event.event_scheduler -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:17:30.037 * Looking for test storage... 00:17:30.037 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event/scheduler 00:17:30.037 20:16:25 event.event_scheduler -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:17:30.037 20:16:25 event.event_scheduler -- common/autotest_common.sh@1681 -- # lcov --version 00:17:30.037 20:16:25 event.event_scheduler -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:17:30.037 20:16:25 event.event_scheduler -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:17:30.037 20:16:25 event.event_scheduler -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:30.037 20:16:25 event.event_scheduler -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:30.037 20:16:25 event.event_scheduler -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:30.037 20:16:25 event.event_scheduler -- scripts/common.sh@336 -- # IFS=.-: 00:17:30.037 20:16:25 event.event_scheduler -- scripts/common.sh@336 -- # read -ra ver1 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@337 -- # IFS=.-: 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@337 -- # read -ra ver2 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@338 -- # local 'op=<' 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@340 -- # ver1_l=2 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@341 -- # ver2_l=1 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@344 -- # case "$op" in 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@345 -- # : 1 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@365 -- # decimal 1 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@353 -- # local d=1 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@355 -- # echo 1 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@365 -- # ver1[v]=1 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@366 -- # decimal 2 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@353 -- # local d=2 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@355 -- # echo 2 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@366 -- # ver2[v]=2 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:30.038 20:16:25 event.event_scheduler -- scripts/common.sh@368 -- # return 0 00:17:30.038 20:16:25 event.event_scheduler -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:30.038 20:16:25 event.event_scheduler -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:17:30.038 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:30.038 --rc genhtml_branch_coverage=1 00:17:30.038 --rc genhtml_function_coverage=1 00:17:30.038 --rc genhtml_legend=1 00:17:30.038 --rc geninfo_all_blocks=1 00:17:30.038 --rc geninfo_unexecuted_blocks=1 00:17:30.038 00:17:30.038 ' 00:17:30.038 20:16:25 event.event_scheduler -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:17:30.038 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:30.038 --rc genhtml_branch_coverage=1 00:17:30.038 --rc genhtml_function_coverage=1 00:17:30.038 --rc genhtml_legend=1 00:17:30.038 --rc geninfo_all_blocks=1 00:17:30.038 --rc geninfo_unexecuted_blocks=1 00:17:30.038 00:17:30.038 ' 00:17:30.038 20:16:25 event.event_scheduler -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:17:30.038 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:30.038 --rc genhtml_branch_coverage=1 00:17:30.038 --rc genhtml_function_coverage=1 00:17:30.038 --rc genhtml_legend=1 00:17:30.038 --rc geninfo_all_blocks=1 00:17:30.038 --rc geninfo_unexecuted_blocks=1 00:17:30.038 00:17:30.038 ' 00:17:30.038 20:16:25 event.event_scheduler -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:17:30.038 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:30.038 --rc genhtml_branch_coverage=1 00:17:30.038 --rc genhtml_function_coverage=1 00:17:30.038 --rc genhtml_legend=1 00:17:30.038 --rc geninfo_all_blocks=1 00:17:30.038 --rc geninfo_unexecuted_blocks=1 00:17:30.038 00:17:30.038 ' 00:17:30.038 20:16:25 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:17:30.038 20:16:25 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=58301 00:17:30.038 20:16:25 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:17:30.038 20:16:25 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:17:30.038 20:16:25 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 58301 00:17:30.038 20:16:25 event.event_scheduler -- common/autotest_common.sh@831 -- # '[' -z 58301 ']' 00:17:30.038 20:16:25 event.event_scheduler -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:30.038 20:16:25 event.event_scheduler -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:30.038 20:16:25 event.event_scheduler -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:30.038 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:30.038 20:16:25 event.event_scheduler -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:30.038 20:16:25 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:17:30.297 [2024-10-01 20:16:25.323329] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:17:30.297 [2024-10-01 20:16:25.323792] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58301 ] 00:17:30.297 [2024-10-01 20:16:25.506036] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 4 00:17:30.556 [2024-10-01 20:16:25.792781] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:30.556 [2024-10-01 20:16:25.792900] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:17:30.556 [2024-10-01 20:16:25.793030] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 3 00:17:30.556 [2024-10-01 20:16:25.793590] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:17:31.123 20:16:26 event.event_scheduler -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:31.123 20:16:26 event.event_scheduler -- common/autotest_common.sh@864 -- # return 0 00:17:31.123 20:16:26 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:17:31.123 20:16:26 event.event_scheduler -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.123 20:16:26 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:17:31.123 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:17:31.123 POWER: Cannot set governor of lcore 0 to userspace 00:17:31.123 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:17:31.123 POWER: Cannot set governor of lcore 0 to performance 00:17:31.123 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:17:31.123 POWER: Cannot set governor of lcore 0 to userspace 00:17:31.123 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:17:31.123 POWER: Cannot set governor of lcore 0 to userspace 00:17:31.123 GUEST_CHANNEL: Opening channel '/dev/virtio-ports/virtio.serial.port.poweragent.0' for lcore 0 00:17:31.123 GUEST_CHANNEL: Unable to connect to '/dev/virtio-ports/virtio.serial.port.poweragent.0' with error No such file or directory 00:17:31.123 POWER: Unable to set Power Management Environment for lcore 0 00:17:31.123 [2024-10-01 20:16:26.295831] dpdk_governor.c: 130:_init_core: *ERROR*: Failed to initialize on core0 00:17:31.123 [2024-10-01 20:16:26.295861] dpdk_governor.c: 191:_init: *ERROR*: Failed to initialize on core0 00:17:31.123 [2024-10-01 20:16:26.295876] scheduler_dynamic.c: 280:init: *NOTICE*: Unable to initialize dpdk governor 00:17:31.123 [2024-10-01 20:16:26.295905] scheduler_dynamic.c: 427:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:17:31.123 [2024-10-01 20:16:26.295920] scheduler_dynamic.c: 429:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:17:31.123 [2024-10-01 20:16:26.295935] scheduler_dynamic.c: 431:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:17:31.123 20:16:26 event.event_scheduler -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:31.123 20:16:26 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:17:31.123 20:16:26 event.event_scheduler -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.123 20:16:26 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:17:31.691 [2024-10-01 20:16:26.903150] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:17:31.691 20:16:26 event.event_scheduler -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:31.691 20:16:26 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:17:31.691 20:16:26 event.event_scheduler -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:17:31.691 20:16:26 event.event_scheduler -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:31.691 20:16:26 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:17:31.691 ************************************ 00:17:31.691 START TEST scheduler_create_thread 00:17:31.691 ************************************ 00:17:31.691 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1125 -- # scheduler_create_thread 00:17:31.691 20:16:26 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:17:31.691 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.691 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:17:31.691 2 00:17:31.691 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:31.691 20:16:26 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:17:31.691 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.691 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:17:31.691 3 00:17:31.691 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:31.691 20:16:26 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:17:31.691 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.691 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:17:31.950 4 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:17:31.950 5 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:17:31.950 6 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:17:31.950 7 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:17:31.950 8 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:17:31.950 9 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:17:31.950 10 00:17:31.950 20:16:26 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:31.950 20:16:27 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:17:31.950 20:16:27 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.950 20:16:27 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:17:31.950 20:16:27 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:31.950 20:16:27 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:17:31.950 20:16:27 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:17:31.950 20:16:27 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.950 20:16:27 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:17:31.950 20:16:27 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:31.951 20:16:27 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:17:31.951 20:16:27 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:31.951 20:16:27 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:17:33.325 20:16:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:33.325 20:16:28 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:17:33.325 20:16:28 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:17:33.325 20:16:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:33.325 20:16:28 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:17:34.698 ************************************ 00:17:34.698 END TEST scheduler_create_thread 00:17:34.698 ************************************ 00:17:34.698 20:16:29 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:34.698 00:17:34.698 real 0m2.621s 00:17:34.698 user 0m0.019s 00:17:34.698 sys 0m0.006s 00:17:34.698 20:16:29 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:34.698 20:16:29 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:17:34.698 20:16:29 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:17:34.698 20:16:29 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 58301 00:17:34.698 20:16:29 event.event_scheduler -- common/autotest_common.sh@950 -- # '[' -z 58301 ']' 00:17:34.698 20:16:29 event.event_scheduler -- common/autotest_common.sh@954 -- # kill -0 58301 00:17:34.698 20:16:29 event.event_scheduler -- common/autotest_common.sh@955 -- # uname 00:17:34.698 20:16:29 event.event_scheduler -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:17:34.698 20:16:29 event.event_scheduler -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 58301 00:17:34.698 killing process with pid 58301 00:17:34.698 20:16:29 event.event_scheduler -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:17:34.699 20:16:29 event.event_scheduler -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:17:34.699 20:16:29 event.event_scheduler -- common/autotest_common.sh@968 -- # echo 'killing process with pid 58301' 00:17:34.699 20:16:29 event.event_scheduler -- common/autotest_common.sh@969 -- # kill 58301 00:17:34.699 20:16:29 event.event_scheduler -- common/autotest_common.sh@974 -- # wait 58301 00:17:34.957 [2024-10-01 20:16:30.018095] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:17:36.884 ************************************ 00:17:36.884 END TEST event_scheduler 00:17:36.884 ************************************ 00:17:36.884 00:17:36.884 real 0m6.866s 00:17:36.884 user 0m12.675s 00:17:36.884 sys 0m0.656s 00:17:36.884 20:16:31 event.event_scheduler -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:36.884 20:16:31 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:17:36.884 20:16:31 event -- event/event.sh@51 -- # modprobe -n nbd 00:17:36.884 20:16:31 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:17:36.884 20:16:31 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:17:36.884 20:16:31 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:36.884 20:16:31 event -- common/autotest_common.sh@10 -- # set +x 00:17:36.884 ************************************ 00:17:36.884 START TEST app_repeat 00:17:36.884 ************************************ 00:17:36.884 20:16:31 event.app_repeat -- common/autotest_common.sh@1125 -- # app_repeat_test 00:17:36.884 20:16:31 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:36.884 20:16:31 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:36.884 20:16:31 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:17:36.884 20:16:31 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:17:36.884 20:16:31 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:17:36.884 20:16:31 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:17:36.884 20:16:31 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:17:36.884 Process app_repeat pid: 58424 00:17:36.884 spdk_app_start Round 0 00:17:36.884 20:16:31 event.app_repeat -- event/event.sh@19 -- # repeat_pid=58424 00:17:36.884 20:16:31 event.app_repeat -- event/event.sh@18 -- # /home/vagrant/spdk_repo/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:17:36.884 20:16:31 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:17:36.884 20:16:31 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 58424' 00:17:36.884 20:16:31 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:17:36.884 20:16:31 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:17:36.884 20:16:31 event.app_repeat -- event/event.sh@25 -- # waitforlisten 58424 /var/tmp/spdk-nbd.sock 00:17:36.884 20:16:31 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 58424 ']' 00:17:36.884 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:17:36.884 20:16:31 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:17:36.884 20:16:31 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:36.884 20:16:31 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:17:36.884 20:16:31 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:36.884 20:16:31 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:17:36.884 [2024-10-01 20:16:32.017135] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:17:36.884 [2024-10-01 20:16:32.017339] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58424 ] 00:17:37.142 [2024-10-01 20:16:32.192334] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:37.400 [2024-10-01 20:16:32.439686] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:37.400 [2024-10-01 20:16:32.439694] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:17:37.966 20:16:33 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:37.966 20:16:33 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:17:37.966 20:16:33 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:17:38.533 Malloc0 00:17:38.533 20:16:33 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:17:38.792 Malloc1 00:17:38.792 20:16:33 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:17:38.792 20:16:33 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:38.792 20:16:33 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:17:38.792 20:16:33 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:17:38.792 20:16:33 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:38.792 20:16:33 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:17:38.792 20:16:33 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:17:38.792 20:16:33 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:38.792 20:16:33 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:17:38.792 20:16:33 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:17:38.792 20:16:33 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:38.792 20:16:33 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:17:38.792 20:16:33 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:17:38.792 20:16:33 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:17:38.792 20:16:33 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:38.792 20:16:33 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:17:39.051 /dev/nbd0 00:17:39.051 20:16:34 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:17:39.051 20:16:34 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:17:39.051 20:16:34 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:17:39.051 20:16:34 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:17:39.051 20:16:34 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:17:39.051 20:16:34 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:17:39.051 20:16:34 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:17:39.051 20:16:34 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:17:39.051 20:16:34 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:17:39.051 20:16:34 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:17:39.051 20:16:34 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:17:39.051 1+0 records in 00:17:39.051 1+0 records out 00:17:39.051 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000315908 s, 13.0 MB/s 00:17:39.051 20:16:34 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:17:39.051 20:16:34 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:17:39.051 20:16:34 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:17:39.051 20:16:34 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:17:39.051 20:16:34 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:17:39.051 20:16:34 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:39.051 20:16:34 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:39.051 20:16:34 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:17:39.310 /dev/nbd1 00:17:39.310 20:16:34 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:17:39.310 20:16:34 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:17:39.310 20:16:34 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:17:39.310 20:16:34 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:17:39.310 20:16:34 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:17:39.310 20:16:34 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:17:39.310 20:16:34 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:17:39.310 20:16:34 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:17:39.310 20:16:34 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:17:39.310 20:16:34 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:17:39.310 20:16:34 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:17:39.310 1+0 records in 00:17:39.310 1+0 records out 00:17:39.310 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000383082 s, 10.7 MB/s 00:17:39.310 20:16:34 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:17:39.310 20:16:34 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:17:39.310 20:16:34 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:17:39.310 20:16:34 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:17:39.310 20:16:34 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:17:39.310 20:16:34 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:39.310 20:16:34 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:39.310 20:16:34 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:39.310 20:16:34 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:39.310 20:16:34 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:17:39.947 { 00:17:39.947 "nbd_device": "/dev/nbd0", 00:17:39.947 "bdev_name": "Malloc0" 00:17:39.947 }, 00:17:39.947 { 00:17:39.947 "nbd_device": "/dev/nbd1", 00:17:39.947 "bdev_name": "Malloc1" 00:17:39.947 } 00:17:39.947 ]' 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:17:39.947 { 00:17:39.947 "nbd_device": "/dev/nbd0", 00:17:39.947 "bdev_name": "Malloc0" 00:17:39.947 }, 00:17:39.947 { 00:17:39.947 "nbd_device": "/dev/nbd1", 00:17:39.947 "bdev_name": "Malloc1" 00:17:39.947 } 00:17:39.947 ]' 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:17:39.947 /dev/nbd1' 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:17:39.947 /dev/nbd1' 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:17:39.947 256+0 records in 00:17:39.947 256+0 records out 00:17:39.947 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0075463 s, 139 MB/s 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:17:39.947 256+0 records in 00:17:39.947 256+0 records out 00:17:39.947 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0327963 s, 32.0 MB/s 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:17:39.947 20:16:34 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:17:39.947 256+0 records in 00:17:39.947 256+0 records out 00:17:39.947 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0388602 s, 27.0 MB/s 00:17:39.947 20:16:35 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:17:39.947 20:16:35 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:39.947 20:16:35 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:17:39.947 20:16:35 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:17:39.947 20:16:35 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:17:39.947 20:16:35 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:17:39.947 20:16:35 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:17:39.947 20:16:35 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:17:39.947 20:16:35 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:17:39.947 20:16:35 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:17:39.947 20:16:35 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:17:39.948 20:16:35 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:17:39.948 20:16:35 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:17:39.948 20:16:35 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:39.948 20:16:35 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:39.948 20:16:35 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:39.948 20:16:35 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:17:39.948 20:16:35 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:39.948 20:16:35 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:17:40.207 20:16:35 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:40.207 20:16:35 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:40.207 20:16:35 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:40.207 20:16:35 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:40.207 20:16:35 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:40.207 20:16:35 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:40.207 20:16:35 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:17:40.207 20:16:35 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:17:40.207 20:16:35 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:40.207 20:16:35 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:17:40.465 20:16:35 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:17:40.465 20:16:35 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:17:40.465 20:16:35 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:17:40.465 20:16:35 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:40.465 20:16:35 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:40.465 20:16:35 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:17:40.465 20:16:35 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:17:40.465 20:16:35 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:17:40.465 20:16:35 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:40.465 20:16:35 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:40.465 20:16:35 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:41.033 20:16:36 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:17:41.033 20:16:36 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:17:41.033 20:16:36 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:41.033 20:16:36 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:17:41.033 20:16:36 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:17:41.033 20:16:36 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:41.033 20:16:36 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:17:41.033 20:16:36 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:17:41.033 20:16:36 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:17:41.033 20:16:36 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:17:41.033 20:16:36 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:17:41.033 20:16:36 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:17:41.033 20:16:36 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:17:41.600 20:16:36 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:17:43.504 [2024-10-01 20:16:38.531151] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:43.788 [2024-10-01 20:16:38.765051] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:17:43.788 [2024-10-01 20:16:38.765065] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:43.788 [2024-10-01 20:16:38.957317] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:17:43.788 [2024-10-01 20:16:38.957399] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:17:44.762 spdk_app_start Round 1 00:17:44.762 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:17:44.762 20:16:39 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:17:44.762 20:16:39 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:17:44.762 20:16:39 event.app_repeat -- event/event.sh@25 -- # waitforlisten 58424 /var/tmp/spdk-nbd.sock 00:17:44.762 20:16:39 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 58424 ']' 00:17:44.762 20:16:39 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:17:44.762 20:16:39 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:44.762 20:16:39 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:17:44.762 20:16:39 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:44.762 20:16:39 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:17:44.762 20:16:39 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:44.762 20:16:39 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:17:44.762 20:16:39 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:17:45.328 Malloc0 00:17:45.328 20:16:40 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:17:45.586 Malloc1 00:17:45.586 20:16:40 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:17:45.586 20:16:40 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:45.586 20:16:40 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:17:45.586 20:16:40 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:17:45.586 20:16:40 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:45.586 20:16:40 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:17:45.586 20:16:40 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:17:45.586 20:16:40 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:45.586 20:16:40 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:17:45.586 20:16:40 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:17:45.586 20:16:40 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:45.586 20:16:40 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:17:45.586 20:16:40 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:17:45.586 20:16:40 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:17:45.586 20:16:40 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:45.586 20:16:40 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:17:45.845 /dev/nbd0 00:17:45.845 20:16:41 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:17:45.845 20:16:41 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:17:45.845 20:16:41 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:17:45.845 20:16:41 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:17:45.845 20:16:41 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:17:45.845 20:16:41 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:17:45.845 20:16:41 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:17:45.845 20:16:41 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:17:45.845 20:16:41 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:17:45.845 20:16:41 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:17:45.845 20:16:41 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:17:45.845 1+0 records in 00:17:45.845 1+0 records out 00:17:45.845 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000390207 s, 10.5 MB/s 00:17:45.845 20:16:41 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:17:45.845 20:16:41 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:17:45.845 20:16:41 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:17:45.845 20:16:41 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:17:45.845 20:16:41 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:17:45.845 20:16:41 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:45.845 20:16:41 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:45.845 20:16:41 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:17:46.412 /dev/nbd1 00:17:46.412 20:16:41 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:17:46.412 20:16:41 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:17:46.412 20:16:41 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:17:46.412 20:16:41 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:17:46.412 20:16:41 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:17:46.412 20:16:41 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:17:46.412 20:16:41 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:17:46.412 20:16:41 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:17:46.412 20:16:41 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:17:46.412 20:16:41 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:17:46.412 20:16:41 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:17:46.412 1+0 records in 00:17:46.412 1+0 records out 00:17:46.412 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000331524 s, 12.4 MB/s 00:17:46.412 20:16:41 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:17:46.412 20:16:41 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:17:46.412 20:16:41 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:17:46.412 20:16:41 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:17:46.412 20:16:41 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:17:46.412 20:16:41 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:46.412 20:16:41 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:46.412 20:16:41 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:46.412 20:16:41 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:46.412 20:16:41 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:46.670 20:16:41 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:17:46.670 { 00:17:46.670 "nbd_device": "/dev/nbd0", 00:17:46.670 "bdev_name": "Malloc0" 00:17:46.670 }, 00:17:46.670 { 00:17:46.670 "nbd_device": "/dev/nbd1", 00:17:46.670 "bdev_name": "Malloc1" 00:17:46.670 } 00:17:46.670 ]' 00:17:46.670 20:16:41 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:17:46.670 { 00:17:46.670 "nbd_device": "/dev/nbd0", 00:17:46.670 "bdev_name": "Malloc0" 00:17:46.670 }, 00:17:46.670 { 00:17:46.670 "nbd_device": "/dev/nbd1", 00:17:46.670 "bdev_name": "Malloc1" 00:17:46.670 } 00:17:46.670 ]' 00:17:46.670 20:16:41 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:46.670 20:16:41 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:17:46.670 /dev/nbd1' 00:17:46.670 20:16:41 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:17:46.670 /dev/nbd1' 00:17:46.670 20:16:41 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:46.670 20:16:41 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:17:46.670 20:16:41 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:17:46.670 20:16:41 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:17:46.670 20:16:41 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:17:46.670 20:16:41 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:17:46.670 20:16:41 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:46.670 20:16:41 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:17:46.670 20:16:41 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:17:46.671 20:16:41 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:17:46.671 20:16:41 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:17:46.671 20:16:41 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:17:46.671 256+0 records in 00:17:46.671 256+0 records out 00:17:46.671 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00883387 s, 119 MB/s 00:17:46.671 20:16:41 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:17:46.671 20:16:41 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:17:46.671 256+0 records in 00:17:46.671 256+0 records out 00:17:46.671 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0323204 s, 32.4 MB/s 00:17:46.671 20:16:41 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:17:46.671 20:16:41 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:17:46.930 256+0 records in 00:17:46.930 256+0 records out 00:17:46.930 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.037392 s, 28.0 MB/s 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:46.930 20:16:41 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:17:47.189 20:16:42 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:47.189 20:16:42 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:47.189 20:16:42 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:47.189 20:16:42 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:47.189 20:16:42 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:47.189 20:16:42 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:47.189 20:16:42 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:17:47.189 20:16:42 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:17:47.189 20:16:42 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:47.189 20:16:42 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:17:47.447 20:16:42 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:17:47.447 20:16:42 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:17:47.447 20:16:42 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:17:47.447 20:16:42 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:47.447 20:16:42 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:47.447 20:16:42 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:17:47.447 20:16:42 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:17:47.447 20:16:42 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:17:47.447 20:16:42 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:47.447 20:16:42 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:47.447 20:16:42 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:47.706 20:16:42 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:17:47.706 20:16:42 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:47.706 20:16:42 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:17:47.965 20:16:42 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:17:47.965 20:16:42 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:17:47.965 20:16:42 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:47.965 20:16:42 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:17:47.965 20:16:42 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:17:47.965 20:16:42 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:17:47.965 20:16:42 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:17:47.965 20:16:42 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:17:47.965 20:16:42 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:17:47.965 20:16:42 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:17:48.224 20:16:43 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:17:50.125 [2024-10-01 20:16:45.313794] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:50.384 [2024-10-01 20:16:45.548661] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:50.384 [2024-10-01 20:16:45.548661] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:17:50.643 [2024-10-01 20:16:45.740811] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:17:50.643 [2024-10-01 20:16:45.740952] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:17:51.209 spdk_app_start Round 2 00:17:51.210 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:17:51.210 20:16:46 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:17:51.210 20:16:46 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:17:51.210 20:16:46 event.app_repeat -- event/event.sh@25 -- # waitforlisten 58424 /var/tmp/spdk-nbd.sock 00:17:51.210 20:16:46 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 58424 ']' 00:17:51.210 20:16:46 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:17:51.210 20:16:46 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:51.210 20:16:46 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:17:51.210 20:16:46 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:51.210 20:16:46 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:17:51.468 20:16:46 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:51.468 20:16:46 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:17:51.468 20:16:46 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:17:52.041 Malloc0 00:17:52.041 20:16:47 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:17:52.299 Malloc1 00:17:52.299 20:16:47 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:17:52.299 20:16:47 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:52.299 20:16:47 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:17:52.299 20:16:47 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:17:52.299 20:16:47 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:52.299 20:16:47 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:17:52.299 20:16:47 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:17:52.299 20:16:47 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:52.299 20:16:47 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:17:52.299 20:16:47 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:17:52.299 20:16:47 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:52.299 20:16:47 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:17:52.299 20:16:47 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:17:52.299 20:16:47 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:17:52.299 20:16:47 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:52.299 20:16:47 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:17:52.557 /dev/nbd0 00:17:52.557 20:16:47 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:17:52.557 20:16:47 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:17:52.557 20:16:47 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:17:52.557 20:16:47 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:17:52.557 20:16:47 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:17:52.557 20:16:47 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:17:52.557 20:16:47 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:17:52.557 20:16:47 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:17:52.557 20:16:47 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:17:52.557 20:16:47 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:17:52.557 20:16:47 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:17:52.557 1+0 records in 00:17:52.557 1+0 records out 00:17:52.557 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000262061 s, 15.6 MB/s 00:17:52.557 20:16:47 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:17:52.557 20:16:47 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:17:52.557 20:16:47 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:17:52.557 20:16:47 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:17:52.557 20:16:47 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:17:52.557 20:16:47 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:52.557 20:16:47 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:52.557 20:16:47 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:17:52.816 /dev/nbd1 00:17:52.816 20:16:48 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:17:53.074 20:16:48 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:17:53.074 20:16:48 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:17:53.074 20:16:48 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:17:53.074 20:16:48 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:17:53.074 20:16:48 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:17:53.074 20:16:48 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:17:53.074 20:16:48 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:17:53.074 20:16:48 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:17:53.074 20:16:48 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:17:53.074 20:16:48 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:17:53.074 1+0 records in 00:17:53.074 1+0 records out 00:17:53.074 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000418954 s, 9.8 MB/s 00:17:53.074 20:16:48 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:17:53.074 20:16:48 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:17:53.074 20:16:48 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:17:53.074 20:16:48 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:17:53.074 20:16:48 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:17:53.074 20:16:48 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:53.074 20:16:48 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:17:53.074 20:16:48 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:53.075 20:16:48 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:53.075 20:16:48 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:17:53.333 { 00:17:53.333 "nbd_device": "/dev/nbd0", 00:17:53.333 "bdev_name": "Malloc0" 00:17:53.333 }, 00:17:53.333 { 00:17:53.333 "nbd_device": "/dev/nbd1", 00:17:53.333 "bdev_name": "Malloc1" 00:17:53.333 } 00:17:53.333 ]' 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:17:53.333 { 00:17:53.333 "nbd_device": "/dev/nbd0", 00:17:53.333 "bdev_name": "Malloc0" 00:17:53.333 }, 00:17:53.333 { 00:17:53.333 "nbd_device": "/dev/nbd1", 00:17:53.333 "bdev_name": "Malloc1" 00:17:53.333 } 00:17:53.333 ]' 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:17:53.333 /dev/nbd1' 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:17:53.333 /dev/nbd1' 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:17:53.333 256+0 records in 00:17:53.333 256+0 records out 00:17:53.333 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00681235 s, 154 MB/s 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:17:53.333 256+0 records in 00:17:53.333 256+0 records out 00:17:53.333 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0316403 s, 33.1 MB/s 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:17:53.333 256+0 records in 00:17:53.333 256+0 records out 00:17:53.333 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0362285 s, 28.9 MB/s 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:53.333 20:16:48 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:17:53.592 20:16:48 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:53.592 20:16:48 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:53.592 20:16:48 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:53.592 20:16:48 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:53.592 20:16:48 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:53.592 20:16:48 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:53.592 20:16:48 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:17:53.592 20:16:48 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:17:53.592 20:16:48 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:53.592 20:16:48 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:17:54.160 20:16:49 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:17:54.160 20:16:49 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:17:54.160 20:16:49 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:17:54.160 20:16:49 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:54.160 20:16:49 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:54.160 20:16:49 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:17:54.160 20:16:49 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:17:54.160 20:16:49 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:17:54.160 20:16:49 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:54.160 20:16:49 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:54.160 20:16:49 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:54.160 20:16:49 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:17:54.160 20:16:49 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:17:54.160 20:16:49 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:54.419 20:16:49 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:17:54.419 20:16:49 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:17:54.419 20:16:49 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:54.419 20:16:49 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:17:54.419 20:16:49 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:17:54.419 20:16:49 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:17:54.419 20:16:49 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:17:54.419 20:16:49 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:17:54.419 20:16:49 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:17:54.419 20:16:49 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:17:54.678 20:16:49 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:17:56.581 [2024-10-01 20:16:51.819048] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:56.840 [2024-10-01 20:16:52.062350] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:17:56.840 [2024-10-01 20:16:52.062361] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:57.100 [2024-10-01 20:16:52.260147] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:17:57.100 [2024-10-01 20:16:52.260224] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:17:58.036 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:17:58.036 20:16:52 event.app_repeat -- event/event.sh@38 -- # waitforlisten 58424 /var/tmp/spdk-nbd.sock 00:17:58.036 20:16:52 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 58424 ']' 00:17:58.036 20:16:52 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:17:58.036 20:16:52 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:58.036 20:16:52 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:17:58.036 20:16:52 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:58.036 20:16:52 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:17:58.036 20:16:53 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:58.036 20:16:53 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:17:58.036 20:16:53 event.app_repeat -- event/event.sh@39 -- # killprocess 58424 00:17:58.036 20:16:53 event.app_repeat -- common/autotest_common.sh@950 -- # '[' -z 58424 ']' 00:17:58.036 20:16:53 event.app_repeat -- common/autotest_common.sh@954 -- # kill -0 58424 00:17:58.036 20:16:53 event.app_repeat -- common/autotest_common.sh@955 -- # uname 00:17:58.036 20:16:53 event.app_repeat -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:17:58.036 20:16:53 event.app_repeat -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 58424 00:17:58.294 killing process with pid 58424 00:17:58.294 20:16:53 event.app_repeat -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:17:58.294 20:16:53 event.app_repeat -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:17:58.294 20:16:53 event.app_repeat -- common/autotest_common.sh@968 -- # echo 'killing process with pid 58424' 00:17:58.294 20:16:53 event.app_repeat -- common/autotest_common.sh@969 -- # kill 58424 00:17:58.294 20:16:53 event.app_repeat -- common/autotest_common.sh@974 -- # wait 58424 00:18:00.193 spdk_app_start is called in Round 0. 00:18:00.193 Shutdown signal received, stop current app iteration 00:18:00.193 Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 reinitialization... 00:18:00.193 spdk_app_start is called in Round 1. 00:18:00.193 Shutdown signal received, stop current app iteration 00:18:00.193 Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 reinitialization... 00:18:00.193 spdk_app_start is called in Round 2. 00:18:00.193 Shutdown signal received, stop current app iteration 00:18:00.193 Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 reinitialization... 00:18:00.193 spdk_app_start is called in Round 3. 00:18:00.193 Shutdown signal received, stop current app iteration 00:18:00.193 20:16:55 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:18:00.193 20:16:55 event.app_repeat -- event/event.sh@42 -- # return 0 00:18:00.193 00:18:00.193 real 0m23.140s 00:18:00.193 user 0m49.812s 00:18:00.193 sys 0m3.604s 00:18:00.193 20:16:55 event.app_repeat -- common/autotest_common.sh@1126 -- # xtrace_disable 00:18:00.193 20:16:55 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:18:00.193 ************************************ 00:18:00.193 END TEST app_repeat 00:18:00.193 ************************************ 00:18:00.193 20:16:55 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:18:00.193 20:16:55 event -- event/event.sh@55 -- # run_test cpu_locks /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:18:00.193 20:16:55 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:18:00.193 20:16:55 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:18:00.193 20:16:55 event -- common/autotest_common.sh@10 -- # set +x 00:18:00.193 ************************************ 00:18:00.193 START TEST cpu_locks 00:18:00.193 ************************************ 00:18:00.193 20:16:55 event.cpu_locks -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:18:00.193 * Looking for test storage... 00:18:00.193 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:18:00.193 20:16:55 event.cpu_locks -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:18:00.193 20:16:55 event.cpu_locks -- common/autotest_common.sh@1681 -- # lcov --version 00:18:00.193 20:16:55 event.cpu_locks -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:18:00.193 20:16:55 event.cpu_locks -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@333 -- # local ver1 ver1_l 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@334 -- # local ver2 ver2_l 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@336 -- # IFS=.-: 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@336 -- # read -ra ver1 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@337 -- # IFS=.-: 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@337 -- # read -ra ver2 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@338 -- # local 'op=<' 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@340 -- # ver1_l=2 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@341 -- # ver2_l=1 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@344 -- # case "$op" in 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@345 -- # : 1 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@364 -- # (( v = 0 )) 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@365 -- # decimal 1 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@353 -- # local d=1 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@355 -- # echo 1 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@365 -- # ver1[v]=1 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@366 -- # decimal 2 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@353 -- # local d=2 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@355 -- # echo 2 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@366 -- # ver2[v]=2 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:18:00.193 20:16:55 event.cpu_locks -- scripts/common.sh@368 -- # return 0 00:18:00.193 20:16:55 event.cpu_locks -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:18:00.193 20:16:55 event.cpu_locks -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:18:00.193 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:00.193 --rc genhtml_branch_coverage=1 00:18:00.193 --rc genhtml_function_coverage=1 00:18:00.193 --rc genhtml_legend=1 00:18:00.193 --rc geninfo_all_blocks=1 00:18:00.193 --rc geninfo_unexecuted_blocks=1 00:18:00.193 00:18:00.193 ' 00:18:00.193 20:16:55 event.cpu_locks -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:18:00.193 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:00.193 --rc genhtml_branch_coverage=1 00:18:00.193 --rc genhtml_function_coverage=1 00:18:00.193 --rc genhtml_legend=1 00:18:00.193 --rc geninfo_all_blocks=1 00:18:00.193 --rc geninfo_unexecuted_blocks=1 00:18:00.193 00:18:00.193 ' 00:18:00.193 20:16:55 event.cpu_locks -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:18:00.193 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:00.193 --rc genhtml_branch_coverage=1 00:18:00.193 --rc genhtml_function_coverage=1 00:18:00.193 --rc genhtml_legend=1 00:18:00.193 --rc geninfo_all_blocks=1 00:18:00.193 --rc geninfo_unexecuted_blocks=1 00:18:00.193 00:18:00.193 ' 00:18:00.193 20:16:55 event.cpu_locks -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:18:00.193 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:00.193 --rc genhtml_branch_coverage=1 00:18:00.193 --rc genhtml_function_coverage=1 00:18:00.193 --rc genhtml_legend=1 00:18:00.193 --rc geninfo_all_blocks=1 00:18:00.193 --rc geninfo_unexecuted_blocks=1 00:18:00.193 00:18:00.193 ' 00:18:00.193 20:16:55 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:18:00.193 20:16:55 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:18:00.193 20:16:55 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:18:00.193 20:16:55 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:18:00.193 20:16:55 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:18:00.193 20:16:55 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:18:00.193 20:16:55 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:18:00.193 ************************************ 00:18:00.193 START TEST default_locks 00:18:00.193 ************************************ 00:18:00.193 20:16:55 event.cpu_locks.default_locks -- common/autotest_common.sh@1125 -- # default_locks 00:18:00.193 20:16:55 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=58920 00:18:00.194 20:16:55 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:18:00.194 20:16:55 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 58920 00:18:00.194 20:16:55 event.cpu_locks.default_locks -- common/autotest_common.sh@831 -- # '[' -z 58920 ']' 00:18:00.194 20:16:55 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:00.194 20:16:55 event.cpu_locks.default_locks -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:00.194 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:00.194 20:16:55 event.cpu_locks.default_locks -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:00.194 20:16:55 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:00.194 20:16:55 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:18:00.452 [2024-10-01 20:16:55.455649] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:18:00.452 [2024-10-01 20:16:55.455843] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid58920 ] 00:18:00.452 [2024-10-01 20:16:55.620305] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:00.711 [2024-10-01 20:16:55.869603] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:18:02.086 20:16:57 event.cpu_locks.default_locks -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:02.086 20:16:57 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # return 0 00:18:02.086 20:16:57 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 58920 00:18:02.086 20:16:57 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 58920 00:18:02.086 20:16:57 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:18:02.652 20:16:57 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 58920 00:18:02.652 20:16:57 event.cpu_locks.default_locks -- common/autotest_common.sh@950 -- # '[' -z 58920 ']' 00:18:02.652 20:16:57 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # kill -0 58920 00:18:02.652 20:16:57 event.cpu_locks.default_locks -- common/autotest_common.sh@955 -- # uname 00:18:02.652 20:16:57 event.cpu_locks.default_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:18:02.652 20:16:57 event.cpu_locks.default_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 58920 00:18:02.652 20:16:57 event.cpu_locks.default_locks -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:18:02.652 killing process with pid 58920 00:18:02.652 20:16:57 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:18:02.652 20:16:57 event.cpu_locks.default_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 58920' 00:18:02.652 20:16:57 event.cpu_locks.default_locks -- common/autotest_common.sh@969 -- # kill 58920 00:18:02.652 20:16:57 event.cpu_locks.default_locks -- common/autotest_common.sh@974 -- # wait 58920 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 58920 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@650 -- # local es=0 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 58920 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@653 -- # waitforlisten 58920 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@831 -- # '[' -z 58920 ']' 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:05.934 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:18:05.934 ERROR: process (pid: 58920) is no longer running 00:18:05.934 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 846: kill: (58920) - No such process 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # return 1 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@653 -- # es=1 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:18:05.934 20:17:00 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:18:05.935 20:17:00 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:18:05.935 20:17:00 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:18:05.935 00:18:05.935 real 0m5.537s 00:18:05.935 user 0m5.449s 00:18:05.935 sys 0m1.062s 00:18:05.935 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:18:05.935 ************************************ 00:18:05.935 END TEST default_locks 00:18:05.935 ************************************ 00:18:05.935 20:17:00 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:18:05.935 20:17:00 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:18:05.935 20:17:00 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:18:05.935 20:17:00 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:18:05.935 20:17:00 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:18:05.935 ************************************ 00:18:05.935 START TEST default_locks_via_rpc 00:18:05.935 ************************************ 00:18:05.935 20:17:00 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1125 -- # default_locks_via_rpc 00:18:05.935 20:17:00 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=59007 00:18:05.935 20:17:00 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 59007 00:18:05.935 20:17:00 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:18:05.935 20:17:00 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 59007 ']' 00:18:05.935 20:17:00 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:05.935 20:17:00 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:05.935 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:05.935 20:17:00 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:05.935 20:17:00 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:05.935 20:17:00 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:18:05.935 [2024-10-01 20:17:01.040726] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:18:05.935 [2024-10-01 20:17:01.040897] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid59007 ] 00:18:06.191 [2024-10-01 20:17:01.217694] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:06.448 [2024-10-01 20:17:01.464449] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 59007 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 59007 00:18:07.460 20:17:02 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:18:08.396 20:17:03 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 59007 00:18:08.396 20:17:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@950 -- # '[' -z 59007 ']' 00:18:08.396 20:17:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # kill -0 59007 00:18:08.396 20:17:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@955 -- # uname 00:18:08.396 20:17:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:18:08.396 20:17:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 59007 00:18:08.396 killing process with pid 59007 00:18:08.396 20:17:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:18:08.396 20:17:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:18:08.396 20:17:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 59007' 00:18:08.396 20:17:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@969 -- # kill 59007 00:18:08.396 20:17:03 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@974 -- # wait 59007 00:18:11.681 ************************************ 00:18:11.681 END TEST default_locks_via_rpc 00:18:11.681 ************************************ 00:18:11.681 00:18:11.681 real 0m5.513s 00:18:11.681 user 0m5.469s 00:18:11.681 sys 0m1.040s 00:18:11.681 20:17:06 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:18:11.681 20:17:06 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:18:11.681 20:17:06 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:18:11.681 20:17:06 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:18:11.681 20:17:06 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:18:11.681 20:17:06 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:18:11.681 ************************************ 00:18:11.681 START TEST non_locking_app_on_locked_coremask 00:18:11.681 ************************************ 00:18:11.681 20:17:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1125 -- # non_locking_app_on_locked_coremask 00:18:11.681 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:11.681 20:17:06 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=59093 00:18:11.681 20:17:06 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:18:11.681 20:17:06 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 59093 /var/tmp/spdk.sock 00:18:11.681 20:17:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 59093 ']' 00:18:11.681 20:17:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:11.681 20:17:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:11.681 20:17:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:11.681 20:17:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:11.681 20:17:06 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:18:11.681 [2024-10-01 20:17:06.605917] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:18:11.681 [2024-10-01 20:17:06.606168] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid59093 ] 00:18:11.681 [2024-10-01 20:17:06.775677] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:11.939 [2024-10-01 20:17:07.022953] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:18:13.322 20:17:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:13.322 20:17:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:18:13.322 20:17:08 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=59120 00:18:13.322 20:17:08 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:18:13.322 20:17:08 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 59120 /var/tmp/spdk2.sock 00:18:13.322 20:17:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 59120 ']' 00:18:13.322 20:17:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:18:13.322 20:17:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:13.322 20:17:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:18:13.322 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:18:13.322 20:17:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:13.322 20:17:08 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:18:13.322 [2024-10-01 20:17:08.304030] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:18:13.322 [2024-10-01 20:17:08.304208] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid59120 ] 00:18:13.322 [2024-10-01 20:17:08.485604] app.c: 914:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:18:13.322 [2024-10-01 20:17:08.485753] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:13.892 [2024-10-01 20:17:08.992771] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:18:16.434 20:17:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:16.435 20:17:11 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:18:16.435 20:17:11 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 59093 00:18:16.435 20:17:11 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 59093 00:18:16.435 20:17:11 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:18:17.810 20:17:12 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 59093 00:18:17.810 20:17:12 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 59093 ']' 00:18:17.810 20:17:12 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 59093 00:18:17.810 20:17:12 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:18:17.810 20:17:12 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:18:17.810 20:17:12 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 59093 00:18:17.810 killing process with pid 59093 00:18:17.810 20:17:12 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:18:17.810 20:17:12 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:18:17.810 20:17:12 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 59093' 00:18:17.810 20:17:12 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 59093 00:18:17.810 20:17:12 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 59093 00:18:24.376 20:17:18 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 59120 00:18:24.376 20:17:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 59120 ']' 00:18:24.376 20:17:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 59120 00:18:24.376 20:17:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:18:24.376 20:17:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:18:24.376 20:17:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 59120 00:18:24.376 killing process with pid 59120 00:18:24.376 20:17:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:18:24.376 20:17:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:18:24.376 20:17:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 59120' 00:18:24.376 20:17:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 59120 00:18:24.376 20:17:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 59120 00:18:26.920 00:18:26.920 real 0m15.294s 00:18:26.920 user 0m15.719s 00:18:26.920 sys 0m2.193s 00:18:26.920 20:17:21 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:18:26.920 20:17:21 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:18:26.920 ************************************ 00:18:26.920 END TEST non_locking_app_on_locked_coremask 00:18:26.920 ************************************ 00:18:26.920 20:17:21 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:18:26.920 20:17:21 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:18:26.920 20:17:21 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:18:26.920 20:17:21 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:18:26.920 ************************************ 00:18:26.920 START TEST locking_app_on_unlocked_coremask 00:18:26.920 ************************************ 00:18:26.920 20:17:21 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1125 -- # locking_app_on_unlocked_coremask 00:18:26.920 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:26.920 20:17:21 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=59301 00:18:26.920 20:17:21 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:18:26.920 20:17:21 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 59301 /var/tmp/spdk.sock 00:18:26.920 20:17:21 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@831 -- # '[' -z 59301 ']' 00:18:26.920 20:17:21 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:26.921 20:17:21 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:26.921 20:17:21 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:26.921 20:17:21 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:26.921 20:17:21 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:18:26.921 [2024-10-01 20:17:21.959399] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:18:26.921 [2024-10-01 20:17:21.959797] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid59301 ] 00:18:26.921 [2024-10-01 20:17:22.127500] app.c: 914:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:18:26.921 [2024-10-01 20:17:22.127936] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:27.179 [2024-10-01 20:17:22.368148] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:18:28.554 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:18:28.554 20:17:23 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:28.554 20:17:23 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # return 0 00:18:28.554 20:17:23 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=59323 00:18:28.554 20:17:23 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 59323 /var/tmp/spdk2.sock 00:18:28.554 20:17:23 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@831 -- # '[' -z 59323 ']' 00:18:28.554 20:17:23 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:18:28.554 20:17:23 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:28.554 20:17:23 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:18:28.554 20:17:23 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:28.554 20:17:23 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:18:28.554 20:17:23 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:18:28.554 [2024-10-01 20:17:23.675408] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:18:28.554 [2024-10-01 20:17:23.675606] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid59323 ] 00:18:28.812 [2024-10-01 20:17:23.873783] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:29.379 [2024-10-01 20:17:24.366924] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:18:31.907 20:17:26 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:31.907 20:17:26 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # return 0 00:18:31.907 20:17:26 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 59323 00:18:31.907 20:17:26 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:18:31.907 20:17:26 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 59323 00:18:32.890 20:17:28 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 59301 00:18:32.890 20:17:28 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@950 -- # '[' -z 59301 ']' 00:18:32.890 20:17:28 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # kill -0 59301 00:18:32.890 20:17:28 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # uname 00:18:32.890 20:17:28 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:18:32.890 20:17:28 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 59301 00:18:32.890 20:17:28 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:18:32.890 20:17:28 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:18:32.890 killing process with pid 59301 00:18:32.890 20:17:28 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 59301' 00:18:32.890 20:17:28 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@969 -- # kill 59301 00:18:32.890 20:17:28 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@974 -- # wait 59301 00:18:39.452 20:17:34 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 59323 00:18:39.452 20:17:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@950 -- # '[' -z 59323 ']' 00:18:39.452 20:17:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # kill -0 59323 00:18:39.452 20:17:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # uname 00:18:39.452 20:17:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:18:39.452 20:17:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 59323 00:18:39.452 killing process with pid 59323 00:18:39.452 20:17:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:18:39.452 20:17:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:18:39.452 20:17:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 59323' 00:18:39.452 20:17:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@969 -- # kill 59323 00:18:39.452 20:17:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@974 -- # wait 59323 00:18:41.984 ************************************ 00:18:41.984 END TEST locking_app_on_unlocked_coremask 00:18:41.984 ************************************ 00:18:41.984 00:18:41.984 real 0m15.261s 00:18:41.984 user 0m15.710s 00:18:41.984 sys 0m2.154s 00:18:41.984 20:17:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:18:41.984 20:17:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:18:41.984 20:17:37 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:18:41.984 20:17:37 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:18:41.984 20:17:37 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:18:41.984 20:17:37 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:18:41.984 ************************************ 00:18:41.984 START TEST locking_app_on_locked_coremask 00:18:41.984 ************************************ 00:18:41.984 20:17:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1125 -- # locking_app_on_locked_coremask 00:18:41.984 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:41.984 20:17:37 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=59505 00:18:41.984 20:17:37 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 59505 /var/tmp/spdk.sock 00:18:41.984 20:17:37 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:18:41.984 20:17:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 59505 ']' 00:18:41.984 20:17:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:41.984 20:17:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:41.984 20:17:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:41.984 20:17:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:41.984 20:17:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:18:42.243 [2024-10-01 20:17:37.271745] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:18:42.243 [2024-10-01 20:17:37.272374] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid59505 ] 00:18:42.243 [2024-10-01 20:17:37.436668] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:42.501 [2024-10-01 20:17:37.701739] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=59532 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 59532 /var/tmp/spdk2.sock 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@650 -- # local es=0 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 59532 /var/tmp/spdk2.sock 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@653 -- # waitforlisten 59532 /var/tmp/spdk2.sock 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 59532 ']' 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:18:43.877 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:43.877 20:17:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:18:43.877 [2024-10-01 20:17:39.052691] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:18:43.877 [2024-10-01 20:17:39.052901] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid59532 ] 00:18:44.136 [2024-10-01 20:17:39.237350] app.c: 779:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 59505 has claimed it. 00:18:44.136 [2024-10-01 20:17:39.237442] app.c: 910:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:18:44.703 ERROR: process (pid: 59532) is no longer running 00:18:44.703 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 846: kill: (59532) - No such process 00:18:44.703 20:17:39 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:44.703 20:17:39 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 1 00:18:44.703 20:17:39 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@653 -- # es=1 00:18:44.703 20:17:39 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:18:44.703 20:17:39 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:18:44.703 20:17:39 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:18:44.703 20:17:39 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 59505 00:18:44.703 20:17:39 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 59505 00:18:44.703 20:17:39 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:18:45.270 20:17:40 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 59505 00:18:45.270 20:17:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 59505 ']' 00:18:45.270 20:17:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 59505 00:18:45.270 20:17:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:18:45.270 20:17:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:18:45.270 20:17:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 59505 00:18:45.270 killing process with pid 59505 00:18:45.270 20:17:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:18:45.270 20:17:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:18:45.270 20:17:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 59505' 00:18:45.270 20:17:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 59505 00:18:45.270 20:17:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 59505 00:18:48.621 00:18:48.621 real 0m6.171s 00:18:48.621 user 0m6.369s 00:18:48.621 sys 0m1.212s 00:18:48.621 20:17:43 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:18:48.621 ************************************ 00:18:48.621 END TEST locking_app_on_locked_coremask 00:18:48.621 ************************************ 00:18:48.621 20:17:43 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:18:48.621 20:17:43 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:18:48.621 20:17:43 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:18:48.621 20:17:43 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:18:48.621 20:17:43 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:18:48.621 ************************************ 00:18:48.621 START TEST locking_overlapped_coremask 00:18:48.621 ************************************ 00:18:48.621 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:48.621 20:17:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1125 -- # locking_overlapped_coremask 00:18:48.621 20:17:43 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=59607 00:18:48.621 20:17:43 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 59607 /var/tmp/spdk.sock 00:18:48.621 20:17:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@831 -- # '[' -z 59607 ']' 00:18:48.621 20:17:43 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 00:18:48.621 20:17:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:48.621 20:17:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:48.621 20:17:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:48.621 20:17:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:48.621 20:17:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:18:48.621 [2024-10-01 20:17:43.524745] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:18:48.621 [2024-10-01 20:17:43.525254] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid59607 ] 00:18:48.621 [2024-10-01 20:17:43.701442] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 3 00:18:48.879 [2024-10-01 20:17:43.984919] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:18:48.880 [2024-10-01 20:17:43.984998] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:18:48.880 [2024-10-01 20:17:43.985017] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # return 0 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=59636 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 59636 /var/tmp/spdk2.sock 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@650 -- # local es=0 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 59636 /var/tmp/spdk2.sock 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@653 -- # waitforlisten 59636 /var/tmp/spdk2.sock 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@831 -- # '[' -z 59636 ']' 00:18:50.255 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:50.255 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:18:50.255 [2024-10-01 20:17:45.350289] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:18:50.255 [2024-10-01 20:17:45.350937] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid59636 ] 00:18:50.513 [2024-10-01 20:17:45.553259] app.c: 779:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 59607 has claimed it. 00:18:50.513 [2024-10-01 20:17:45.553370] app.c: 910:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:18:50.772 ERROR: process (pid: 59636) is no longer running 00:18:50.772 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 846: kill: (59636) - No such process 00:18:50.772 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:50.772 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # return 1 00:18:50.772 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@653 -- # es=1 00:18:50.772 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:18:50.772 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:18:50.772 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:18:50.772 20:17:45 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:18:50.772 20:17:45 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:18:50.772 20:17:45 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:18:50.772 20:17:45 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:18:50.772 20:17:45 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 59607 00:18:50.772 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@950 -- # '[' -z 59607 ']' 00:18:50.772 20:17:45 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # kill -0 59607 00:18:50.772 20:17:46 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@955 -- # uname 00:18:50.772 20:17:46 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:18:50.772 20:17:46 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 59607 00:18:51.030 20:17:46 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:18:51.031 20:17:46 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:18:51.031 20:17:46 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 59607' 00:18:51.031 killing process with pid 59607 00:18:51.031 20:17:46 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@969 -- # kill 59607 00:18:51.031 20:17:46 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@974 -- # wait 59607 00:18:54.386 00:18:54.386 real 0m5.740s 00:18:54.386 user 0m15.156s 00:18:54.386 sys 0m0.876s 00:18:54.386 20:17:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:18:54.386 20:17:49 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:18:54.386 ************************************ 00:18:54.386 END TEST locking_overlapped_coremask 00:18:54.386 ************************************ 00:18:54.386 20:17:49 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:18:54.386 20:17:49 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:18:54.386 20:17:49 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:18:54.386 20:17:49 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:18:54.386 ************************************ 00:18:54.386 START TEST locking_overlapped_coremask_via_rpc 00:18:54.386 ************************************ 00:18:54.386 20:17:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1125 -- # locking_overlapped_coremask_via_rpc 00:18:54.386 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:54.386 20:17:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=59711 00:18:54.386 20:17:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:18:54.386 20:17:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 59711 /var/tmp/spdk.sock 00:18:54.386 20:17:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 59711 ']' 00:18:54.386 20:17:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:54.386 20:17:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:54.386 20:17:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:54.386 20:17:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:54.386 20:17:49 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:18:54.386 [2024-10-01 20:17:49.299354] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:18:54.386 [2024-10-01 20:17:49.299530] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid59711 ] 00:18:54.386 [2024-10-01 20:17:49.467242] app.c: 914:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:18:54.386 [2024-10-01 20:17:49.467314] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 3 00:18:54.644 [2024-10-01 20:17:49.721236] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:18:54.644 [2024-10-01 20:17:49.721286] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:18:54.644 [2024-10-01 20:17:49.721308] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:18:56.018 20:17:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:56.018 20:17:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:18:56.018 20:17:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=59735 00:18:56.018 20:17:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:18:56.018 20:17:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 59735 /var/tmp/spdk2.sock 00:18:56.018 20:17:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 59735 ']' 00:18:56.018 20:17:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:18:56.018 20:17:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:56.018 20:17:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:18:56.018 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:18:56.018 20:17:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:56.018 20:17:50 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:18:56.018 [2024-10-01 20:17:50.995147] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:18:56.018 [2024-10-01 20:17:50.995394] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid59735 ] 00:18:56.018 [2024-10-01 20:17:51.188636] app.c: 914:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:18:56.018 [2024-10-01 20:17:51.188727] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 3 00:18:56.585 [2024-10-01 20:17:51.694807] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 3 00:18:56.585 [2024-10-01 20:17:51.694867] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:18:56.585 [2024-10-01 20:17:51.694877] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 4 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@650 -- # local es=0 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@653 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:18:59.114 [2024-10-01 20:17:53.965004] app.c: 779:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 59711 has claimed it. 00:18:59.114 request: 00:18:59.114 { 00:18:59.114 "method": "framework_enable_cpumask_locks", 00:18:59.114 "req_id": 1 00:18:59.114 } 00:18:59.114 Got JSON-RPC error response 00:18:59.114 response: 00:18:59.114 { 00:18:59.114 "code": -32603, 00:18:59.114 "message": "Failed to claim CPU core: 2" 00:18:59.114 } 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@653 -- # es=1 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 59711 /var/tmp/spdk.sock 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 59711 ']' 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:59.114 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:59.114 20:17:53 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:18:59.114 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:18:59.114 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:59.114 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:18:59.114 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 59735 /var/tmp/spdk2.sock 00:18:59.114 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 59735 ']' 00:18:59.114 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:18:59.114 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:59.114 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:18:59.114 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:59.114 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:18:59.373 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:59.373 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:18:59.373 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:18:59.373 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:18:59.373 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:18:59.374 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:18:59.374 00:18:59.374 real 0m5.439s 00:18:59.374 user 0m2.001s 00:18:59.374 sys 0m0.279s 00:18:59.374 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:18:59.374 20:17:54 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:18:59.374 ************************************ 00:18:59.374 END TEST locking_overlapped_coremask_via_rpc 00:18:59.374 ************************************ 00:18:59.632 20:17:54 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:18:59.632 20:17:54 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 59711 ]] 00:18:59.632 20:17:54 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 59711 00:18:59.632 20:17:54 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 59711 ']' 00:18:59.632 20:17:54 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 59711 00:18:59.632 20:17:54 event.cpu_locks -- common/autotest_common.sh@955 -- # uname 00:18:59.632 20:17:54 event.cpu_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:18:59.632 20:17:54 event.cpu_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 59711 00:18:59.632 killing process with pid 59711 00:18:59.632 20:17:54 event.cpu_locks -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:18:59.632 20:17:54 event.cpu_locks -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:18:59.632 20:17:54 event.cpu_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 59711' 00:18:59.632 20:17:54 event.cpu_locks -- common/autotest_common.sh@969 -- # kill 59711 00:18:59.632 20:17:54 event.cpu_locks -- common/autotest_common.sh@974 -- # wait 59711 00:19:02.971 20:17:57 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 59735 ]] 00:19:02.971 20:17:57 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 59735 00:19:02.971 20:17:57 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 59735 ']' 00:19:02.971 20:17:57 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 59735 00:19:02.971 20:17:57 event.cpu_locks -- common/autotest_common.sh@955 -- # uname 00:19:02.971 20:17:57 event.cpu_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:02.971 20:17:57 event.cpu_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 59735 00:19:02.971 killing process with pid 59735 00:19:02.971 20:17:57 event.cpu_locks -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:19:02.971 20:17:57 event.cpu_locks -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:19:02.971 20:17:57 event.cpu_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 59735' 00:19:02.971 20:17:57 event.cpu_locks -- common/autotest_common.sh@969 -- # kill 59735 00:19:02.971 20:17:57 event.cpu_locks -- common/autotest_common.sh@974 -- # wait 59735 00:19:06.262 20:18:00 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:19:06.262 20:18:00 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:19:06.262 20:18:00 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 59711 ]] 00:19:06.262 20:18:00 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 59711 00:19:06.262 20:18:00 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 59711 ']' 00:19:06.262 Process with pid 59711 is not found 00:19:06.262 20:18:00 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 59711 00:19:06.262 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 954: kill: (59711) - No such process 00:19:06.262 20:18:00 event.cpu_locks -- common/autotest_common.sh@977 -- # echo 'Process with pid 59711 is not found' 00:19:06.262 20:18:00 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 59735 ]] 00:19:06.262 20:18:00 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 59735 00:19:06.262 Process with pid 59735 is not found 00:19:06.262 20:18:00 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 59735 ']' 00:19:06.262 20:18:00 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 59735 00:19:06.262 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 954: kill: (59735) - No such process 00:19:06.263 20:18:00 event.cpu_locks -- common/autotest_common.sh@977 -- # echo 'Process with pid 59735 is not found' 00:19:06.263 20:18:00 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:19:06.263 00:19:06.263 real 1m5.703s 00:19:06.263 user 1m49.310s 00:19:06.263 sys 0m10.333s 00:19:06.263 20:18:00 event.cpu_locks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:06.263 20:18:00 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:19:06.263 ************************************ 00:19:06.263 END TEST cpu_locks 00:19:06.263 ************************************ 00:19:06.263 ************************************ 00:19:06.263 END TEST event 00:19:06.263 ************************************ 00:19:06.263 00:19:06.263 real 1m41.868s 00:19:06.263 user 2m59.920s 00:19:06.263 sys 0m15.253s 00:19:06.263 20:18:00 event -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:06.263 20:18:00 event -- common/autotest_common.sh@10 -- # set +x 00:19:06.263 20:18:00 -- spdk/autotest.sh@169 -- # run_test thread /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:19:06.263 20:18:00 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:19:06.263 20:18:00 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:06.263 20:18:00 -- common/autotest_common.sh@10 -- # set +x 00:19:06.263 ************************************ 00:19:06.263 START TEST thread 00:19:06.263 ************************************ 00:19:06.263 20:18:00 thread -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:19:06.263 * Looking for test storage... 00:19:06.263 * Found test storage at /home/vagrant/spdk_repo/spdk/test/thread 00:19:06.263 20:18:01 thread -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:19:06.263 20:18:01 thread -- common/autotest_common.sh@1681 -- # lcov --version 00:19:06.263 20:18:01 thread -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:19:06.263 20:18:01 thread -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:19:06.263 20:18:01 thread -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:19:06.263 20:18:01 thread -- scripts/common.sh@333 -- # local ver1 ver1_l 00:19:06.263 20:18:01 thread -- scripts/common.sh@334 -- # local ver2 ver2_l 00:19:06.263 20:18:01 thread -- scripts/common.sh@336 -- # IFS=.-: 00:19:06.263 20:18:01 thread -- scripts/common.sh@336 -- # read -ra ver1 00:19:06.263 20:18:01 thread -- scripts/common.sh@337 -- # IFS=.-: 00:19:06.263 20:18:01 thread -- scripts/common.sh@337 -- # read -ra ver2 00:19:06.263 20:18:01 thread -- scripts/common.sh@338 -- # local 'op=<' 00:19:06.263 20:18:01 thread -- scripts/common.sh@340 -- # ver1_l=2 00:19:06.263 20:18:01 thread -- scripts/common.sh@341 -- # ver2_l=1 00:19:06.263 20:18:01 thread -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:19:06.263 20:18:01 thread -- scripts/common.sh@344 -- # case "$op" in 00:19:06.263 20:18:01 thread -- scripts/common.sh@345 -- # : 1 00:19:06.263 20:18:01 thread -- scripts/common.sh@364 -- # (( v = 0 )) 00:19:06.263 20:18:01 thread -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:06.263 20:18:01 thread -- scripts/common.sh@365 -- # decimal 1 00:19:06.263 20:18:01 thread -- scripts/common.sh@353 -- # local d=1 00:19:06.263 20:18:01 thread -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:19:06.263 20:18:01 thread -- scripts/common.sh@355 -- # echo 1 00:19:06.263 20:18:01 thread -- scripts/common.sh@365 -- # ver1[v]=1 00:19:06.263 20:18:01 thread -- scripts/common.sh@366 -- # decimal 2 00:19:06.263 20:18:01 thread -- scripts/common.sh@353 -- # local d=2 00:19:06.263 20:18:01 thread -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:19:06.263 20:18:01 thread -- scripts/common.sh@355 -- # echo 2 00:19:06.263 20:18:01 thread -- scripts/common.sh@366 -- # ver2[v]=2 00:19:06.263 20:18:01 thread -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:19:06.263 20:18:01 thread -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:19:06.263 20:18:01 thread -- scripts/common.sh@368 -- # return 0 00:19:06.263 20:18:01 thread -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:19:06.263 20:18:01 thread -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:19:06.263 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:06.263 --rc genhtml_branch_coverage=1 00:19:06.263 --rc genhtml_function_coverage=1 00:19:06.263 --rc genhtml_legend=1 00:19:06.263 --rc geninfo_all_blocks=1 00:19:06.263 --rc geninfo_unexecuted_blocks=1 00:19:06.263 00:19:06.263 ' 00:19:06.263 20:18:01 thread -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:19:06.263 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:06.263 --rc genhtml_branch_coverage=1 00:19:06.263 --rc genhtml_function_coverage=1 00:19:06.263 --rc genhtml_legend=1 00:19:06.263 --rc geninfo_all_blocks=1 00:19:06.263 --rc geninfo_unexecuted_blocks=1 00:19:06.263 00:19:06.263 ' 00:19:06.263 20:18:01 thread -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:19:06.263 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:06.263 --rc genhtml_branch_coverage=1 00:19:06.263 --rc genhtml_function_coverage=1 00:19:06.263 --rc genhtml_legend=1 00:19:06.263 --rc geninfo_all_blocks=1 00:19:06.263 --rc geninfo_unexecuted_blocks=1 00:19:06.263 00:19:06.263 ' 00:19:06.263 20:18:01 thread -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:19:06.263 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:06.263 --rc genhtml_branch_coverage=1 00:19:06.263 --rc genhtml_function_coverage=1 00:19:06.263 --rc genhtml_legend=1 00:19:06.263 --rc geninfo_all_blocks=1 00:19:06.263 --rc geninfo_unexecuted_blocks=1 00:19:06.263 00:19:06.263 ' 00:19:06.263 20:18:01 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:19:06.263 20:18:01 thread -- common/autotest_common.sh@1101 -- # '[' 8 -le 1 ']' 00:19:06.263 20:18:01 thread -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:06.263 20:18:01 thread -- common/autotest_common.sh@10 -- # set +x 00:19:06.263 ************************************ 00:19:06.263 START TEST thread_poller_perf 00:19:06.263 ************************************ 00:19:06.263 20:18:01 thread.thread_poller_perf -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:19:06.263 [2024-10-01 20:18:01.195931] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:19:06.263 [2024-10-01 20:18:01.196421] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid59952 ] 00:19:06.263 [2024-10-01 20:18:01.377344] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:06.522 [2024-10-01 20:18:01.662846] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:19:06.522 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:19:07.899 ====================================== 00:19:07.899 busy:2210981444 (cyc) 00:19:07.899 total_run_count: 302000 00:19:07.899 tsc_hz: 2200000000 (cyc) 00:19:07.899 ====================================== 00:19:07.899 poller_cost: 7321 (cyc), 3327 (nsec) 00:19:07.899 00:19:07.899 ************************************ 00:19:07.899 END TEST thread_poller_perf 00:19:07.899 ************************************ 00:19:07.899 real 0m1.947s 00:19:07.899 user 0m1.701s 00:19:07.899 sys 0m0.131s 00:19:07.899 20:18:03 thread.thread_poller_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:07.899 20:18:03 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:19:07.899 20:18:03 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:19:07.899 20:18:03 thread -- common/autotest_common.sh@1101 -- # '[' 8 -le 1 ']' 00:19:07.899 20:18:03 thread -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:07.899 20:18:03 thread -- common/autotest_common.sh@10 -- # set +x 00:19:07.899 ************************************ 00:19:07.899 START TEST thread_poller_perf 00:19:07.899 ************************************ 00:19:07.899 20:18:03 thread.thread_poller_perf -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:19:08.158 [2024-10-01 20:18:03.202945] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:19:08.158 [2024-10-01 20:18:03.203782] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid59994 ] 00:19:08.158 [2024-10-01 20:18:03.400546] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:08.725 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:19:08.725 [2024-10-01 20:18:03.677010] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:19:10.099 ====================================== 00:19:10.099 busy:2204259014 (cyc) 00:19:10.099 total_run_count: 3685000 00:19:10.099 tsc_hz: 2200000000 (cyc) 00:19:10.099 ====================================== 00:19:10.099 poller_cost: 598 (cyc), 271 (nsec) 00:19:10.099 00:19:10.099 real 0m1.951s 00:19:10.099 user 0m1.697s 00:19:10.099 sys 0m0.139s 00:19:10.099 20:18:05 thread.thread_poller_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:10.099 20:18:05 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:19:10.099 ************************************ 00:19:10.099 END TEST thread_poller_perf 00:19:10.099 ************************************ 00:19:10.099 20:18:05 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:19:10.099 ************************************ 00:19:10.099 END TEST thread 00:19:10.099 ************************************ 00:19:10.099 00:19:10.099 real 0m4.203s 00:19:10.099 user 0m3.546s 00:19:10.099 sys 0m0.424s 00:19:10.099 20:18:05 thread -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:10.099 20:18:05 thread -- common/autotest_common.sh@10 -- # set +x 00:19:10.099 20:18:05 -- spdk/autotest.sh@171 -- # [[ 0 -eq 1 ]] 00:19:10.099 20:18:05 -- spdk/autotest.sh@176 -- # run_test app_cmdline /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:19:10.099 20:18:05 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:19:10.099 20:18:05 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:10.099 20:18:05 -- common/autotest_common.sh@10 -- # set +x 00:19:10.099 ************************************ 00:19:10.099 START TEST app_cmdline 00:19:10.099 ************************************ 00:19:10.099 20:18:05 app_cmdline -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:19:10.099 * Looking for test storage... 00:19:10.099 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:19:10.099 20:18:05 app_cmdline -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:19:10.099 20:18:05 app_cmdline -- common/autotest_common.sh@1681 -- # lcov --version 00:19:10.099 20:18:05 app_cmdline -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:19:10.357 20:18:05 app_cmdline -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@333 -- # local ver1 ver1_l 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@334 -- # local ver2 ver2_l 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@336 -- # IFS=.-: 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@336 -- # read -ra ver1 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@337 -- # IFS=.-: 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@337 -- # read -ra ver2 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@338 -- # local 'op=<' 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@340 -- # ver1_l=2 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@341 -- # ver2_l=1 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@344 -- # case "$op" in 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@345 -- # : 1 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@364 -- # (( v = 0 )) 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@365 -- # decimal 1 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@353 -- # local d=1 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@355 -- # echo 1 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@365 -- # ver1[v]=1 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@366 -- # decimal 2 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@353 -- # local d=2 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@355 -- # echo 2 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@366 -- # ver2[v]=2 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:19:10.357 20:18:05 app_cmdline -- scripts/common.sh@368 -- # return 0 00:19:10.357 20:18:05 app_cmdline -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:19:10.357 20:18:05 app_cmdline -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:19:10.357 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:10.357 --rc genhtml_branch_coverage=1 00:19:10.357 --rc genhtml_function_coverage=1 00:19:10.357 --rc genhtml_legend=1 00:19:10.357 --rc geninfo_all_blocks=1 00:19:10.357 --rc geninfo_unexecuted_blocks=1 00:19:10.357 00:19:10.357 ' 00:19:10.357 20:18:05 app_cmdline -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:19:10.357 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:10.357 --rc genhtml_branch_coverage=1 00:19:10.357 --rc genhtml_function_coverage=1 00:19:10.357 --rc genhtml_legend=1 00:19:10.357 --rc geninfo_all_blocks=1 00:19:10.357 --rc geninfo_unexecuted_blocks=1 00:19:10.357 00:19:10.357 ' 00:19:10.357 20:18:05 app_cmdline -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:19:10.357 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:10.357 --rc genhtml_branch_coverage=1 00:19:10.357 --rc genhtml_function_coverage=1 00:19:10.357 --rc genhtml_legend=1 00:19:10.357 --rc geninfo_all_blocks=1 00:19:10.357 --rc geninfo_unexecuted_blocks=1 00:19:10.357 00:19:10.357 ' 00:19:10.357 20:18:05 app_cmdline -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:19:10.357 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:10.357 --rc genhtml_branch_coverage=1 00:19:10.357 --rc genhtml_function_coverage=1 00:19:10.357 --rc genhtml_legend=1 00:19:10.357 --rc geninfo_all_blocks=1 00:19:10.357 --rc geninfo_unexecuted_blocks=1 00:19:10.357 00:19:10.357 ' 00:19:10.358 20:18:05 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:19:10.358 20:18:05 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=60083 00:19:10.358 20:18:05 app_cmdline -- app/cmdline.sh@16 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:19:10.358 20:18:05 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 60083 00:19:10.358 20:18:05 app_cmdline -- common/autotest_common.sh@831 -- # '[' -z 60083 ']' 00:19:10.358 20:18:05 app_cmdline -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:10.358 20:18:05 app_cmdline -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:10.358 20:18:05 app_cmdline -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:10.358 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:10.358 20:18:05 app_cmdline -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:10.358 20:18:05 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:19:10.358 [2024-10-01 20:18:05.528297] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:19:10.358 [2024-10-01 20:18:05.528785] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid60083 ] 00:19:10.616 [2024-10-01 20:18:05.711924] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:10.875 [2024-10-01 20:18:05.983648] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:19:12.257 20:18:07 app_cmdline -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:12.257 20:18:07 app_cmdline -- common/autotest_common.sh@864 -- # return 0 00:19:12.257 20:18:07 app_cmdline -- app/cmdline.sh@20 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py spdk_get_version 00:19:12.257 { 00:19:12.257 "version": "SPDK v25.01-pre git sha1 1b1c3081e", 00:19:12.257 "fields": { 00:19:12.257 "major": 25, 00:19:12.257 "minor": 1, 00:19:12.257 "patch": 0, 00:19:12.257 "suffix": "-pre", 00:19:12.257 "commit": "1b1c3081e" 00:19:12.257 } 00:19:12.257 } 00:19:12.257 20:18:07 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:19:12.257 20:18:07 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:19:12.257 20:18:07 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:19:12.257 20:18:07 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:19:12.257 20:18:07 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:19:12.257 20:18:07 app_cmdline -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:12.257 20:18:07 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:19:12.257 20:18:07 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:19:12.257 20:18:07 app_cmdline -- app/cmdline.sh@26 -- # sort 00:19:12.257 20:18:07 app_cmdline -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:12.258 20:18:07 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:19:12.258 20:18:07 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:19:12.258 20:18:07 app_cmdline -- app/cmdline.sh@30 -- # NOT /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:19:12.258 20:18:07 app_cmdline -- common/autotest_common.sh@650 -- # local es=0 00:19:12.258 20:18:07 app_cmdline -- common/autotest_common.sh@652 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:19:12.258 20:18:07 app_cmdline -- common/autotest_common.sh@638 -- # local arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:19:12.258 20:18:07 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:19:12.258 20:18:07 app_cmdline -- common/autotest_common.sh@642 -- # type -t /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:19:12.258 20:18:07 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:19:12.258 20:18:07 app_cmdline -- common/autotest_common.sh@644 -- # type -P /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:19:12.258 20:18:07 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:19:12.258 20:18:07 app_cmdline -- common/autotest_common.sh@644 -- # arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:19:12.258 20:18:07 app_cmdline -- common/autotest_common.sh@644 -- # [[ -x /home/vagrant/spdk_repo/spdk/scripts/rpc.py ]] 00:19:12.258 20:18:07 app_cmdline -- common/autotest_common.sh@653 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:19:12.827 request: 00:19:12.827 { 00:19:12.827 "method": "env_dpdk_get_mem_stats", 00:19:12.827 "req_id": 1 00:19:12.827 } 00:19:12.827 Got JSON-RPC error response 00:19:12.827 response: 00:19:12.827 { 00:19:12.827 "code": -32601, 00:19:12.827 "message": "Method not found" 00:19:12.827 } 00:19:12.827 20:18:07 app_cmdline -- common/autotest_common.sh@653 -- # es=1 00:19:12.827 20:18:07 app_cmdline -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:19:12.827 20:18:07 app_cmdline -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:19:12.827 20:18:07 app_cmdline -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:19:12.827 20:18:07 app_cmdline -- app/cmdline.sh@1 -- # killprocess 60083 00:19:12.827 20:18:07 app_cmdline -- common/autotest_common.sh@950 -- # '[' -z 60083 ']' 00:19:12.827 20:18:07 app_cmdline -- common/autotest_common.sh@954 -- # kill -0 60083 00:19:12.827 20:18:07 app_cmdline -- common/autotest_common.sh@955 -- # uname 00:19:12.827 20:18:07 app_cmdline -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:12.827 20:18:07 app_cmdline -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 60083 00:19:12.827 killing process with pid 60083 00:19:12.827 20:18:07 app_cmdline -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:19:12.827 20:18:07 app_cmdline -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:19:12.827 20:18:07 app_cmdline -- common/autotest_common.sh@968 -- # echo 'killing process with pid 60083' 00:19:12.827 20:18:07 app_cmdline -- common/autotest_common.sh@969 -- # kill 60083 00:19:12.827 20:18:07 app_cmdline -- common/autotest_common.sh@974 -- # wait 60083 00:19:16.111 ************************************ 00:19:16.111 END TEST app_cmdline 00:19:16.111 ************************************ 00:19:16.111 00:19:16.111 real 0m5.748s 00:19:16.111 user 0m6.015s 00:19:16.111 sys 0m0.866s 00:19:16.111 20:18:10 app_cmdline -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:16.111 20:18:10 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:19:16.111 20:18:10 -- spdk/autotest.sh@177 -- # run_test version /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:19:16.111 20:18:10 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:19:16.111 20:18:10 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:16.111 20:18:10 -- common/autotest_common.sh@10 -- # set +x 00:19:16.111 ************************************ 00:19:16.111 START TEST version 00:19:16.111 ************************************ 00:19:16.111 20:18:10 version -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:19:16.111 * Looking for test storage... 00:19:16.111 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:19:16.111 20:18:11 version -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:19:16.111 20:18:11 version -- common/autotest_common.sh@1681 -- # lcov --version 00:19:16.111 20:18:11 version -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:19:16.111 20:18:11 version -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:19:16.111 20:18:11 version -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:19:16.111 20:18:11 version -- scripts/common.sh@333 -- # local ver1 ver1_l 00:19:16.111 20:18:11 version -- scripts/common.sh@334 -- # local ver2 ver2_l 00:19:16.111 20:18:11 version -- scripts/common.sh@336 -- # IFS=.-: 00:19:16.111 20:18:11 version -- scripts/common.sh@336 -- # read -ra ver1 00:19:16.111 20:18:11 version -- scripts/common.sh@337 -- # IFS=.-: 00:19:16.111 20:18:11 version -- scripts/common.sh@337 -- # read -ra ver2 00:19:16.111 20:18:11 version -- scripts/common.sh@338 -- # local 'op=<' 00:19:16.111 20:18:11 version -- scripts/common.sh@340 -- # ver1_l=2 00:19:16.111 20:18:11 version -- scripts/common.sh@341 -- # ver2_l=1 00:19:16.111 20:18:11 version -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:19:16.111 20:18:11 version -- scripts/common.sh@344 -- # case "$op" in 00:19:16.111 20:18:11 version -- scripts/common.sh@345 -- # : 1 00:19:16.111 20:18:11 version -- scripts/common.sh@364 -- # (( v = 0 )) 00:19:16.111 20:18:11 version -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:16.111 20:18:11 version -- scripts/common.sh@365 -- # decimal 1 00:19:16.111 20:18:11 version -- scripts/common.sh@353 -- # local d=1 00:19:16.111 20:18:11 version -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:19:16.111 20:18:11 version -- scripts/common.sh@355 -- # echo 1 00:19:16.111 20:18:11 version -- scripts/common.sh@365 -- # ver1[v]=1 00:19:16.111 20:18:11 version -- scripts/common.sh@366 -- # decimal 2 00:19:16.111 20:18:11 version -- scripts/common.sh@353 -- # local d=2 00:19:16.111 20:18:11 version -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:19:16.111 20:18:11 version -- scripts/common.sh@355 -- # echo 2 00:19:16.111 20:18:11 version -- scripts/common.sh@366 -- # ver2[v]=2 00:19:16.111 20:18:11 version -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:19:16.111 20:18:11 version -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:19:16.111 20:18:11 version -- scripts/common.sh@368 -- # return 0 00:19:16.111 20:18:11 version -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:19:16.111 20:18:11 version -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:19:16.111 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:16.111 --rc genhtml_branch_coverage=1 00:19:16.111 --rc genhtml_function_coverage=1 00:19:16.111 --rc genhtml_legend=1 00:19:16.111 --rc geninfo_all_blocks=1 00:19:16.111 --rc geninfo_unexecuted_blocks=1 00:19:16.111 00:19:16.111 ' 00:19:16.111 20:18:11 version -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:19:16.111 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:16.111 --rc genhtml_branch_coverage=1 00:19:16.111 --rc genhtml_function_coverage=1 00:19:16.111 --rc genhtml_legend=1 00:19:16.111 --rc geninfo_all_blocks=1 00:19:16.111 --rc geninfo_unexecuted_blocks=1 00:19:16.111 00:19:16.111 ' 00:19:16.111 20:18:11 version -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:19:16.111 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:16.111 --rc genhtml_branch_coverage=1 00:19:16.111 --rc genhtml_function_coverage=1 00:19:16.111 --rc genhtml_legend=1 00:19:16.111 --rc geninfo_all_blocks=1 00:19:16.111 --rc geninfo_unexecuted_blocks=1 00:19:16.111 00:19:16.111 ' 00:19:16.111 20:18:11 version -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:19:16.111 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:16.111 --rc genhtml_branch_coverage=1 00:19:16.111 --rc genhtml_function_coverage=1 00:19:16.111 --rc genhtml_legend=1 00:19:16.111 --rc geninfo_all_blocks=1 00:19:16.111 --rc geninfo_unexecuted_blocks=1 00:19:16.111 00:19:16.111 ' 00:19:16.111 20:18:11 version -- app/version.sh@17 -- # get_header_version major 00:19:16.111 20:18:11 version -- app/version.sh@14 -- # cut -f2 00:19:16.111 20:18:11 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:19:16.111 20:18:11 version -- app/version.sh@14 -- # tr -d '"' 00:19:16.111 20:18:11 version -- app/version.sh@17 -- # major=25 00:19:16.111 20:18:11 version -- app/version.sh@18 -- # get_header_version minor 00:19:16.111 20:18:11 version -- app/version.sh@14 -- # cut -f2 00:19:16.111 20:18:11 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:19:16.111 20:18:11 version -- app/version.sh@14 -- # tr -d '"' 00:19:16.111 20:18:11 version -- app/version.sh@18 -- # minor=1 00:19:16.112 20:18:11 version -- app/version.sh@19 -- # get_header_version patch 00:19:16.112 20:18:11 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:19:16.112 20:18:11 version -- app/version.sh@14 -- # cut -f2 00:19:16.112 20:18:11 version -- app/version.sh@14 -- # tr -d '"' 00:19:16.112 20:18:11 version -- app/version.sh@19 -- # patch=0 00:19:16.112 20:18:11 version -- app/version.sh@20 -- # get_header_version suffix 00:19:16.112 20:18:11 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:19:16.112 20:18:11 version -- app/version.sh@14 -- # cut -f2 00:19:16.112 20:18:11 version -- app/version.sh@14 -- # tr -d '"' 00:19:16.112 20:18:11 version -- app/version.sh@20 -- # suffix=-pre 00:19:16.112 20:18:11 version -- app/version.sh@22 -- # version=25.1 00:19:16.112 20:18:11 version -- app/version.sh@25 -- # (( patch != 0 )) 00:19:16.112 20:18:11 version -- app/version.sh@28 -- # version=25.1rc0 00:19:16.112 20:18:11 version -- app/version.sh@30 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python 00:19:16.112 20:18:11 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:19:16.112 20:18:11 version -- app/version.sh@30 -- # py_version=25.1rc0 00:19:16.112 20:18:11 version -- app/version.sh@31 -- # [[ 25.1rc0 == \2\5\.\1\r\c\0 ]] 00:19:16.112 ************************************ 00:19:16.112 END TEST version 00:19:16.112 ************************************ 00:19:16.112 00:19:16.112 real 0m0.250s 00:19:16.112 user 0m0.152s 00:19:16.112 sys 0m0.131s 00:19:16.112 20:18:11 version -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:16.112 20:18:11 version -- common/autotest_common.sh@10 -- # set +x 00:19:16.112 20:18:11 -- spdk/autotest.sh@179 -- # '[' 0 -eq 1 ']' 00:19:16.112 20:18:11 -- spdk/autotest.sh@188 -- # [[ 1 -eq 1 ]] 00:19:16.112 20:18:11 -- spdk/autotest.sh@189 -- # run_test bdev_raid /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:19:16.112 20:18:11 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:19:16.112 20:18:11 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:16.112 20:18:11 -- common/autotest_common.sh@10 -- # set +x 00:19:16.112 ************************************ 00:19:16.112 START TEST bdev_raid 00:19:16.112 ************************************ 00:19:16.112 20:18:11 bdev_raid -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:19:16.370 * Looking for test storage... 00:19:16.370 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:19:16.370 20:18:11 bdev_raid -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:19:16.370 20:18:11 bdev_raid -- common/autotest_common.sh@1681 -- # lcov --version 00:19:16.370 20:18:11 bdev_raid -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:19:16.370 20:18:11 bdev_raid -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@336 -- # IFS=.-: 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@336 -- # read -ra ver1 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@337 -- # IFS=.-: 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@337 -- # read -ra ver2 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@338 -- # local 'op=<' 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@340 -- # ver1_l=2 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@341 -- # ver2_l=1 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@344 -- # case "$op" in 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@345 -- # : 1 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@365 -- # decimal 1 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@353 -- # local d=1 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@355 -- # echo 1 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@366 -- # decimal 2 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@353 -- # local d=2 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@355 -- # echo 2 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:19:16.370 20:18:11 bdev_raid -- scripts/common.sh@368 -- # return 0 00:19:16.370 20:18:11 bdev_raid -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:19:16.370 20:18:11 bdev_raid -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:19:16.370 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:16.370 --rc genhtml_branch_coverage=1 00:19:16.370 --rc genhtml_function_coverage=1 00:19:16.370 --rc genhtml_legend=1 00:19:16.370 --rc geninfo_all_blocks=1 00:19:16.370 --rc geninfo_unexecuted_blocks=1 00:19:16.370 00:19:16.370 ' 00:19:16.370 20:18:11 bdev_raid -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:19:16.370 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:16.370 --rc genhtml_branch_coverage=1 00:19:16.370 --rc genhtml_function_coverage=1 00:19:16.370 --rc genhtml_legend=1 00:19:16.370 --rc geninfo_all_blocks=1 00:19:16.370 --rc geninfo_unexecuted_blocks=1 00:19:16.370 00:19:16.370 ' 00:19:16.370 20:18:11 bdev_raid -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:19:16.370 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:16.370 --rc genhtml_branch_coverage=1 00:19:16.370 --rc genhtml_function_coverage=1 00:19:16.370 --rc genhtml_legend=1 00:19:16.370 --rc geninfo_all_blocks=1 00:19:16.370 --rc geninfo_unexecuted_blocks=1 00:19:16.370 00:19:16.370 ' 00:19:16.370 20:18:11 bdev_raid -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:19:16.370 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:16.370 --rc genhtml_branch_coverage=1 00:19:16.370 --rc genhtml_function_coverage=1 00:19:16.370 --rc genhtml_legend=1 00:19:16.370 --rc geninfo_all_blocks=1 00:19:16.370 --rc geninfo_unexecuted_blocks=1 00:19:16.370 00:19:16.370 ' 00:19:16.370 20:18:11 bdev_raid -- bdev/bdev_raid.sh@12 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:19:16.370 20:18:11 bdev_raid -- bdev/nbd_common.sh@6 -- # set -e 00:19:16.370 20:18:11 bdev_raid -- bdev/bdev_raid.sh@14 -- # rpc_py=rpc_cmd 00:19:16.370 20:18:11 bdev_raid -- bdev/bdev_raid.sh@946 -- # mkdir -p /raidtest 00:19:16.370 20:18:11 bdev_raid -- bdev/bdev_raid.sh@947 -- # trap 'cleanup; exit 1' EXIT 00:19:16.370 20:18:11 bdev_raid -- bdev/bdev_raid.sh@949 -- # base_blocklen=512 00:19:16.370 20:18:11 bdev_raid -- bdev/bdev_raid.sh@951 -- # run_test raid1_resize_data_offset_test raid_resize_data_offset_test 00:19:16.370 20:18:11 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:19:16.370 20:18:11 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:16.370 20:18:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:19:16.370 ************************************ 00:19:16.370 START TEST raid1_resize_data_offset_test 00:19:16.370 ************************************ 00:19:16.370 20:18:11 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1125 -- # raid_resize_data_offset_test 00:19:16.370 20:18:11 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@917 -- # raid_pid=60287 00:19:16.370 20:18:11 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@918 -- # echo 'Process raid pid: 60287' 00:19:16.370 Process raid pid: 60287 00:19:16.370 20:18:11 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@919 -- # waitforlisten 60287 00:19:16.370 20:18:11 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@916 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:19:16.370 20:18:11 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@831 -- # '[' -z 60287 ']' 00:19:16.370 20:18:11 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:16.370 20:18:11 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:16.370 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:16.370 20:18:11 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:16.370 20:18:11 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:16.370 20:18:11 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:19:16.629 [2024-10-01 20:18:11.631964] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:19:16.629 [2024-10-01 20:18:11.632216] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:16.629 [2024-10-01 20:18:11.817426] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:16.887 [2024-10-01 20:18:12.101368] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:19:17.145 [2024-10-01 20:18:12.311300] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:17.145 [2024-10-01 20:18:12.311389] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:17.712 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:17.712 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@864 -- # return 0 00:19:17.712 20:18:12 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@922 -- # rpc_cmd bdev_malloc_create -b malloc0 64 512 -o 16 00:19:17.712 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:17.712 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:19:17.712 malloc0 00:19:17.712 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:17.712 20:18:12 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@923 -- # rpc_cmd bdev_malloc_create -b malloc1 64 512 -o 16 00:19:17.712 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:17.712 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:19:17.712 malloc1 00:19:17.712 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:17.712 20:18:12 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@924 -- # rpc_cmd bdev_null_create null0 64 512 00:19:17.712 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:17.712 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:19:17.712 null0 00:19:17.712 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:17.713 20:18:12 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@926 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''malloc0 malloc1 null0'\''' -s 00:19:17.713 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:17.713 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:19:17.713 [2024-10-01 20:18:12.916467] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc0 is claimed 00:19:17.713 [2024-10-01 20:18:12.919144] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:19:17.713 [2024-10-01 20:18:12.919216] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev null0 is claimed 00:19:17.713 [2024-10-01 20:18:12.919427] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:19:17.713 [2024-10-01 20:18:12.919457] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 129024, blocklen 512 00:19:17.713 [2024-10-01 20:18:12.919848] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:19:17.713 [2024-10-01 20:18:12.920094] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:19:17.713 [2024-10-01 20:18:12.920117] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:19:17.713 [2024-10-01 20:18:12.920369] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:17.713 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:17.713 20:18:12 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # rpc_cmd bdev_raid_get_bdevs all 00:19:17.713 20:18:12 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:19:17.713 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:17.713 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:19:17.713 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:17.972 20:18:12 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # (( 2048 == 2048 )) 00:19:17.972 20:18:12 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@931 -- # rpc_cmd bdev_null_delete null0 00:19:17.972 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:17.972 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:19:17.972 [2024-10-01 20:18:12.980545] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: null0 00:19:17.972 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:17.972 20:18:12 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@935 -- # rpc_cmd bdev_malloc_create -b malloc2 512 512 -o 30 00:19:17.972 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:17.972 20:18:12 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:19:18.538 malloc2 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@936 -- # rpc_cmd bdev_raid_add_base_bdev Raid malloc2 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:19:18.538 [2024-10-01 20:18:13.564061] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:19:18.538 [2024-10-01 20:18:13.580455] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:18.538 [2024-10-01 20:18:13.583235] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev Raid 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # rpc_cmd bdev_raid_get_bdevs all 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # (( 2070 == 2070 )) 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@941 -- # killprocess 60287 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@950 -- # '[' -z 60287 ']' 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@954 -- # kill -0 60287 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@955 -- # uname 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 60287 00:19:18.538 killing process with pid 60287 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 60287' 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@969 -- # kill 60287 00:19:18.538 20:18:13 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@974 -- # wait 60287 00:19:18.538 [2024-10-01 20:18:13.676549] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:19:18.538 [2024-10-01 20:18:13.678844] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev Raid: Operation canceled 00:19:18.538 [2024-10-01 20:18:13.678965] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:18.538 [2024-10-01 20:18:13.678993] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: malloc2 00:19:18.538 [2024-10-01 20:18:13.706055] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:19:18.538 [2024-10-01 20:18:13.706630] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:19:18.538 [2024-10-01 20:18:13.706672] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:19:20.465 [2024-10-01 20:18:15.409472] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:19:22.368 20:18:17 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@943 -- # return 0 00:19:22.368 00:19:22.368 real 0m5.671s 00:19:22.368 user 0m5.502s 00:19:22.368 sys 0m0.836s 00:19:22.368 20:18:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:22.368 ************************************ 00:19:22.368 20:18:17 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:19:22.368 END TEST raid1_resize_data_offset_test 00:19:22.368 ************************************ 00:19:22.368 20:18:17 bdev_raid -- bdev/bdev_raid.sh@953 -- # run_test raid0_resize_superblock_test raid_resize_superblock_test 0 00:19:22.368 20:18:17 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:19:22.368 20:18:17 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:22.368 20:18:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:19:22.368 ************************************ 00:19:22.368 START TEST raid0_resize_superblock_test 00:19:22.368 ************************************ 00:19:22.368 20:18:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1125 -- # raid_resize_superblock_test 0 00:19:22.368 20:18:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=0 00:19:22.368 Process raid pid: 60381 00:19:22.368 20:18:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=60381 00:19:22.368 20:18:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:19:22.368 20:18:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 60381' 00:19:22.368 20:18:17 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 60381 00:19:22.368 20:18:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 60381 ']' 00:19:22.368 20:18:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:22.368 20:18:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:22.368 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:22.368 20:18:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:22.368 20:18:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:22.368 20:18:17 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:22.368 [2024-10-01 20:18:17.351523] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:19:22.368 [2024-10-01 20:18:17.351783] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:22.368 [2024-10-01 20:18:17.527869] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:22.642 [2024-10-01 20:18:17.770898] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:19:22.903 [2024-10-01 20:18:17.965964] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:22.903 [2024-10-01 20:18:17.966053] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:23.162 20:18:18 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:23.162 20:18:18 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:19:23.162 20:18:18 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:19:23.162 20:18:18 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:23.162 20:18:18 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:23.730 malloc0 00:19:23.730 20:18:18 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:23.730 20:18:18 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:19:23.730 20:18:18 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:23.730 20:18:18 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:23.730 [2024-10-01 20:18:18.906103] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:19:23.730 [2024-10-01 20:18:18.906219] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:23.730 [2024-10-01 20:18:18.906254] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:19:23.730 [2024-10-01 20:18:18.906273] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:23.730 [2024-10-01 20:18:18.909166] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:23.730 [2024-10-01 20:18:18.909223] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:19:23.730 pt0 00:19:23.730 20:18:18 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:23.730 20:18:18 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:19:23.730 20:18:18 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:23.730 20:18:18 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:23.993 4917392b-efd9-4850-81b8-7a6f2c602dc3 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:23.993 9f259a15-c847-4ae9-864c-88ca5dec4f0f 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:23.993 b040106a-ad26-4e3e-9c22-72c078f4e36f 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@870 -- # rpc_cmd bdev_raid_create -n Raid -r 0 -z 64 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:23.993 [2024-10-01 20:18:19.059591] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 9f259a15-c847-4ae9-864c-88ca5dec4f0f is claimed 00:19:23.993 [2024-10-01 20:18:19.059718] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev b040106a-ad26-4e3e-9c22-72c078f4e36f is claimed 00:19:23.993 [2024-10-01 20:18:19.059960] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:19:23.993 [2024-10-01 20:18:19.059998] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 245760, blocklen 512 00:19:23.993 [2024-10-01 20:18:19.060306] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:19:23.993 [2024-10-01 20:18:19.060569] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:19:23.993 [2024-10-01 20:18:19.060596] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:19:23.993 [2024-10-01 20:18:19.060804] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:23.993 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # rpc_cmd bdev_get_bdevs -b Raid 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # jq '.[].num_blocks' 00:19:23.994 [2024-10-01 20:18:19.184080] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # (( 245760 == 245760 )) 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:23.994 [2024-10-01 20:18:19.236042] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:19:23.994 [2024-10-01 20:18:19.236118] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '9f259a15-c847-4ae9-864c-88ca5dec4f0f' was resized: old size 131072, new size 204800 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:23.994 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:23.994 [2024-10-01 20:18:19.243821] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:19:23.994 [2024-10-01 20:18:19.243854] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'b040106a-ad26-4e3e-9c22-72c078f4e36f' was resized: old size 131072, new size 204800 00:19:23.994 [2024-10-01 20:18:19.243891] bdev_raid.c:2340:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 245760 to 393216 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # rpc_cmd bdev_get_bdevs -b Raid 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # jq '.[].num_blocks' 00:19:24.254 [2024-10-01 20:18:19.364032] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # (( 393216 == 393216 )) 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:24.254 [2024-10-01 20:18:19.411714] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:19:24.254 [2024-10-01 20:18:19.411838] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:19:24.254 [2024-10-01 20:18:19.411866] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:19:24.254 [2024-10-01 20:18:19.411904] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:19:24.254 [2024-10-01 20:18:19.412051] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:19:24.254 [2024-10-01 20:18:19.412113] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:19:24.254 [2024-10-01 20:18:19.412134] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:24.254 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:24.254 [2024-10-01 20:18:19.419667] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:19:24.254 [2024-10-01 20:18:19.419780] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:24.254 [2024-10-01 20:18:19.419815] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008180 00:19:24.254 [2024-10-01 20:18:19.419834] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:24.254 [2024-10-01 20:18:19.423041] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:24.255 [2024-10-01 20:18:19.423141] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:19:24.255 pt0 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:24.255 [2024-10-01 20:18:19.425686] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 9f259a15-c847-4ae9-864c-88ca5dec4f0f 00:19:24.255 [2024-10-01 20:18:19.425805] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 9f259a15-c847-4ae9-864c-88ca5dec4f0f is claimed 00:19:24.255 [2024-10-01 20:18:19.425977] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev b040106a-ad26-4e3e-9c22-72c078f4e36f 00:19:24.255 [2024-10-01 20:18:19.426011] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev b040106a-ad26-4e3e-9c22-72c078f4e36f is claimed 00:19:24.255 [2024-10-01 20:18:19.426217] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev b040106a-ad26-4e3e-9c22-72c078f4e36f (2) smaller than existing raid bdev Raid (3) 00:19:24.255 [2024-10-01 20:18:19.426249] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 9f259a15-c847-4ae9-864c-88ca5dec4f0f: File exists 00:19:24.255 [2024-10-01 20:18:19.426330] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:19:24.255 [2024-10-01 20:18:19.426350] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 393216, blocklen 512 00:19:24.255 [2024-10-01 20:18:19.426738] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:19:24.255 [2024-10-01 20:18:19.426945] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:19:24.255 [2024-10-01 20:18:19.426962] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007b00 00:19:24.255 [2024-10-01 20:18:19.427213] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # rpc_cmd bdev_get_bdevs -b Raid 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # jq '.[].num_blocks' 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:24.255 [2024-10-01 20:18:19.439999] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # (( 393216 == 393216 )) 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 60381 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 60381 ']' 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@954 -- # kill -0 60381 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@955 -- # uname 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:24.255 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 60381 00:19:24.514 killing process with pid 60381 00:19:24.514 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:19:24.514 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:19:24.514 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 60381' 00:19:24.514 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@969 -- # kill 60381 00:19:24.514 [2024-10-01 20:18:19.517780] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:19:24.514 20:18:19 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@974 -- # wait 60381 00:19:24.514 [2024-10-01 20:18:19.517904] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:19:24.514 [2024-10-01 20:18:19.517973] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:19:24.514 [2024-10-01 20:18:19.517989] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Raid, state offline 00:19:25.891 [2024-10-01 20:18:20.796893] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:19:27.794 20:18:22 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:19:27.794 00:19:27.794 real 0m5.295s 00:19:27.794 user 0m5.417s 00:19:27.794 sys 0m0.801s 00:19:27.794 20:18:22 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:27.794 ************************************ 00:19:27.794 END TEST raid0_resize_superblock_test 00:19:27.794 ************************************ 00:19:27.794 20:18:22 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:27.794 20:18:22 bdev_raid -- bdev/bdev_raid.sh@954 -- # run_test raid1_resize_superblock_test raid_resize_superblock_test 1 00:19:27.794 20:18:22 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:19:27.794 20:18:22 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:27.794 20:18:22 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:19:27.794 ************************************ 00:19:27.794 START TEST raid1_resize_superblock_test 00:19:27.794 ************************************ 00:19:27.794 20:18:22 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1125 -- # raid_resize_superblock_test 1 00:19:27.794 20:18:22 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=1 00:19:27.794 20:18:22 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=60486 00:19:27.794 Process raid pid: 60486 00:19:27.794 20:18:22 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 60486' 00:19:27.794 20:18:22 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:19:27.794 20:18:22 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 60486 00:19:27.794 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:27.794 20:18:22 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 60486 ']' 00:19:27.794 20:18:22 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:27.794 20:18:22 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:27.794 20:18:22 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:27.794 20:18:22 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:27.794 20:18:22 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:27.794 [2024-10-01 20:18:22.683607] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:19:27.794 [2024-10-01 20:18:22.684140] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:27.794 [2024-10-01 20:18:22.849341] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:28.056 [2024-10-01 20:18:23.101555] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:19:28.056 [2024-10-01 20:18:23.301967] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:28.056 [2024-10-01 20:18:23.302032] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:28.625 20:18:23 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:28.625 20:18:23 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:19:28.625 20:18:23 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:19:28.625 20:18:23 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:28.625 20:18:23 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.193 malloc0 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.193 [2024-10-01 20:18:24.265710] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:19:29.193 [2024-10-01 20:18:24.266022] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:29.193 [2024-10-01 20:18:24.266072] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:19:29.193 [2024-10-01 20:18:24.266096] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:29.193 [2024-10-01 20:18:24.269137] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:29.193 [2024-10-01 20:18:24.269331] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:19:29.193 pt0 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.193 15aa0f46-8b3b-40b7-8559-20dcd3c5e5b6 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.193 411f698a-192c-4529-a4d1-70eca6c46840 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.193 71c25bc2-028e-4164-b754-8c5b263b2239 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@871 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.193 [2024-10-01 20:18:24.426091] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 411f698a-192c-4529-a4d1-70eca6c46840 is claimed 00:19:29.193 [2024-10-01 20:18:24.426358] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 71c25bc2-028e-4164-b754-8c5b263b2239 is claimed 00:19:29.193 [2024-10-01 20:18:24.426576] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:19:29.193 [2024-10-01 20:18:24.426607] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 122880, blocklen 512 00:19:29.193 [2024-10-01 20:18:24.426962] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005c70 00:19:29.193 [2024-10-01 20:18:24.427216] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:19:29.193 [2024-10-01 20:18:24.427241] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:19:29.193 [2024-10-01 20:18:24.427453] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.193 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # rpc_cmd bdev_get_bdevs -b Raid 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # jq '.[].num_blocks' 00:19:29.453 [2024-10-01 20:18:24.538474] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # (( 122880 == 122880 )) 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.453 [2024-10-01 20:18:24.590471] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:19:29.453 [2024-10-01 20:18:24.590510] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '411f698a-192c-4529-a4d1-70eca6c46840' was resized: old size 131072, new size 204800 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.453 [2024-10-01 20:18:24.598302] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:19:29.453 [2024-10-01 20:18:24.598329] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '71c25bc2-028e-4164-b754-8c5b263b2239' was resized: old size 131072, new size 204800 00:19:29.453 [2024-10-01 20:18:24.598388] bdev_raid.c:2340:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 122880 to 196608 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.453 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.712 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:19:29.712 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:19:29.712 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # rpc_cmd bdev_get_bdevs -b Raid 00:19:29.712 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.712 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.712 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:19:29.712 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # jq '.[].num_blocks' 00:19:29.712 [2024-10-01 20:18:24.718489] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:19:29.712 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.712 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:19:29.712 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # (( 196608 == 196608 )) 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.713 [2024-10-01 20:18:24.770178] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:19:29.713 [2024-10-01 20:18:24.770447] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:19:29.713 [2024-10-01 20:18:24.770541] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:19:29.713 [2024-10-01 20:18:24.770957] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:19:29.713 [2024-10-01 20:18:24.771430] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:19:29.713 [2024-10-01 20:18:24.771677] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:19:29.713 [2024-10-01 20:18:24.771884] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.713 [2024-10-01 20:18:24.782155] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:19:29.713 [2024-10-01 20:18:24.782384] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:19:29.713 [2024-10-01 20:18:24.782477] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008180 00:19:29.713 [2024-10-01 20:18:24.782741] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:19:29.713 [2024-10-01 20:18:24.785951] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:19:29.713 [2024-10-01 20:18:24.786037] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:19:29.713 pt0 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.713 [2024-10-01 20:18:24.788518] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 411f698a-192c-4529-a4d1-70eca6c46840 00:19:29.713 [2024-10-01 20:18:24.788810] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 411f698a-192c-4529-a4d1-70eca6c46840 is claimed 00:19:29.713 [2024-10-01 20:18:24.788977] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 71c25bc2-028e-4164-b754-8c5b263b2239 00:19:29.713 [2024-10-01 20:18:24.789037] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 71c25bc2-028e-4164-b754-8c5b263b2239 is claimed 00:19:29.713 [2024-10-01 20:18:24.789195] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev 71c25bc2-028e-4164-b754-8c5b263b2239 (2) smaller than existing raid bdev Raid (3) 00:19:29.713 [2024-10-01 20:18:24.789229] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 411f698a-192c-4529-a4d1-70eca6c46840: File exists 00:19:29.713 [2024-10-01 20:18:24.789290] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:19:29.713 [2024-10-01 20:18:24.789311] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:19:29.713 [2024-10-01 20:18:24.789701] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:19:29.713 [2024-10-01 20:18:24.789954] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:19:29.713 [2024-10-01 20:18:24.789971] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007b00 00:19:29.713 [2024-10-01 20:18:24.790321] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # rpc_cmd bdev_get_bdevs -b Raid 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # jq '.[].num_blocks' 00:19:29.713 [2024-10-01 20:18:24.802572] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # (( 196608 == 196608 )) 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 60486 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 60486 ']' 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@954 -- # kill -0 60486 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@955 -- # uname 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 60486 00:19:29.713 killing process with pid 60486 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 60486' 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@969 -- # kill 60486 00:19:29.713 [2024-10-01 20:18:24.886475] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:19:29.713 [2024-10-01 20:18:24.886566] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:19:29.713 20:18:24 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@974 -- # wait 60486 00:19:29.713 [2024-10-01 20:18:24.886636] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:19:29.713 [2024-10-01 20:18:24.886669] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Raid, state offline 00:19:31.090 [2024-10-01 20:18:26.184679] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:19:32.993 ************************************ 00:19:32.993 END TEST raid1_resize_superblock_test 00:19:32.993 ************************************ 00:19:32.993 20:18:28 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:19:32.993 00:19:32.993 real 0m5.423s 00:19:32.993 user 0m5.635s 00:19:32.993 sys 0m0.773s 00:19:32.993 20:18:28 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:32.993 20:18:28 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:19:32.993 20:18:28 bdev_raid -- bdev/bdev_raid.sh@956 -- # uname -s 00:19:32.993 20:18:28 bdev_raid -- bdev/bdev_raid.sh@956 -- # '[' Linux = Linux ']' 00:19:32.993 20:18:28 bdev_raid -- bdev/bdev_raid.sh@956 -- # modprobe -n nbd 00:19:32.993 20:18:28 bdev_raid -- bdev/bdev_raid.sh@957 -- # has_nbd=true 00:19:32.993 20:18:28 bdev_raid -- bdev/bdev_raid.sh@958 -- # modprobe nbd 00:19:32.993 20:18:28 bdev_raid -- bdev/bdev_raid.sh@959 -- # run_test raid_function_test_raid0 raid_function_test raid0 00:19:32.993 20:18:28 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:19:32.993 20:18:28 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:32.993 20:18:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:19:32.993 ************************************ 00:19:32.993 START TEST raid_function_test_raid0 00:19:32.993 ************************************ 00:19:32.993 20:18:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1125 -- # raid_function_test raid0 00:19:32.993 20:18:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@64 -- # local raid_level=raid0 00:19:32.993 20:18:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:19:32.993 20:18:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:19:32.993 Process raid pid: 60594 00:19:32.993 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:32.993 20:18:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@69 -- # raid_pid=60594 00:19:32.993 20:18:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:19:32.993 20:18:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 60594' 00:19:32.993 20:18:28 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@71 -- # waitforlisten 60594 00:19:32.993 20:18:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@831 -- # '[' -z 60594 ']' 00:19:32.993 20:18:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:32.993 20:18:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:32.993 20:18:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:32.993 20:18:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:32.993 20:18:28 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:19:32.993 [2024-10-01 20:18:28.205250] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:19:32.994 [2024-10-01 20:18:28.205751] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:33.252 [2024-10-01 20:18:28.389756] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:33.510 [2024-10-01 20:18:28.673218] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:19:33.769 [2024-10-01 20:18:28.877521] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:33.769 [2024-10-01 20:18:28.877625] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:34.334 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:34.334 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@864 -- # return 0 00:19:34.334 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:19:34.334 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:34.334 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:19:34.334 Base_1 00:19:34.334 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:34.334 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:19:34.334 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:34.334 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:19:34.334 Base_2 00:19:34.334 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:34.334 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''Base_1 Base_2'\''' -n raid 00:19:34.334 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:34.334 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:19:34.334 [2024-10-01 20:18:29.399416] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:19:34.334 [2024-10-01 20:18:29.402225] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:19:34.334 [2024-10-01 20:18:29.402519] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:19:34.334 [2024-10-01 20:18:29.402550] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:19:34.334 [2024-10-01 20:18:29.402919] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:19:34.334 [2024-10-01 20:18:29.403152] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:19:34.334 [2024-10-01 20:18:29.403168] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000007780 00:19:34.334 [2024-10-01 20:18:29.403461] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:34.334 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:34.334 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:19:34.334 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:19:34.335 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:34.335 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:19:34.335 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:34.335 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:19:34.335 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:19:34.335 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:19:34.335 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:19:34.335 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:19:34.335 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # local bdev_list 00:19:34.335 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:19:34.335 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # local nbd_list 00:19:34.335 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@12 -- # local i 00:19:34.335 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:19:34.335 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:19:34.335 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:19:34.593 [2024-10-01 20:18:29.711603] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:19:34.593 /dev/nbd0 00:19:34.593 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:19:34.593 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:19:34.593 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:19:34.593 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@869 -- # local i 00:19:34.593 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:19:34.594 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:19:34.594 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:19:34.594 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@873 -- # break 00:19:34.594 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:19:34.594 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:19:34.594 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:19:34.594 1+0 records in 00:19:34.594 1+0 records out 00:19:34.594 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000504986 s, 8.1 MB/s 00:19:34.594 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:19:34.594 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@886 -- # size=4096 00:19:34.594 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:19:34.594 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:19:34.594 20:18:29 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@889 -- # return 0 00:19:34.594 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:19:34.594 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:19:34.594 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:19:34.594 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:19:34.594 20:18:29 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:19:34.852 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:19:34.852 { 00:19:34.852 "nbd_device": "/dev/nbd0", 00:19:34.852 "bdev_name": "raid" 00:19:34.852 } 00:19:34.852 ]' 00:19:34.852 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[ 00:19:34.852 { 00:19:34.852 "nbd_device": "/dev/nbd0", 00:19:34.852 "bdev_name": "raid" 00:19:34.852 } 00:19:34.852 ]' 00:19:34.852 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=1 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 1 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # count=1 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@19 -- # local blksize 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # blksize=512 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:19:35.111 4096+0 records in 00:19:35.111 4096+0 records out 00:19:35.111 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0327839 s, 64.0 MB/s 00:19:35.111 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:19:35.390 4096+0 records in 00:19:35.390 4096+0 records out 00:19:35.390 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.327999 s, 6.4 MB/s 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:19:35.390 128+0 records in 00:19:35.390 128+0 records out 00:19:35.390 65536 bytes (66 kB, 64 KiB) copied, 0.000503937 s, 130 MB/s 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:19:35.390 2035+0 records in 00:19:35.390 2035+0 records out 00:19:35.390 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0128468 s, 81.1 MB/s 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:19:35.390 456+0 records in 00:19:35.390 456+0 records out 00:19:35.390 233472 bytes (233 kB, 228 KiB) copied, 0.00295227 s, 79.1 MB/s 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:19:35.390 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:19:35.663 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:19:35.663 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:19:35.663 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@52 -- # return 0 00:19:35.663 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:19:35.663 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:19:35.663 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:19:35.663 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # local nbd_list 00:19:35.663 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@51 -- # local i 00:19:35.663 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:19:35.663 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:19:35.921 [2024-10-01 20:18:30.953646] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:35.921 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:19:35.921 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:19:35.921 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:19:35.921 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:19:35.921 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:19:35.921 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:19:35.921 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@41 -- # break 00:19:35.921 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@45 -- # return 0 00:19:35.921 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:19:35.921 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:19:35.921 20:18:30 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[]' 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo '' 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # true 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=0 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 0 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # count=0 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@97 -- # killprocess 60594 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@950 -- # '[' -z 60594 ']' 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@954 -- # kill -0 60594 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@955 -- # uname 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 60594 00:19:36.179 killing process with pid 60594 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@968 -- # echo 'killing process with pid 60594' 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@969 -- # kill 60594 00:19:36.179 [2024-10-01 20:18:31.352494] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:19:36.179 20:18:31 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@974 -- # wait 60594 00:19:36.179 [2024-10-01 20:18:31.352634] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:19:36.179 [2024-10-01 20:18:31.352698] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:19:36.180 [2024-10-01 20:18:31.352763] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid, state offline 00:19:36.438 [2024-10-01 20:18:31.541136] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:19:38.340 20:18:33 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@99 -- # return 0 00:19:38.340 00:19:38.340 real 0m5.230s 00:19:38.340 user 0m6.035s 00:19:38.340 sys 0m1.203s 00:19:38.340 20:18:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:38.340 ************************************ 00:19:38.340 END TEST raid_function_test_raid0 00:19:38.340 ************************************ 00:19:38.340 20:18:33 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:19:38.340 20:18:33 bdev_raid -- bdev/bdev_raid.sh@960 -- # run_test raid_function_test_concat raid_function_test concat 00:19:38.340 20:18:33 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:19:38.340 20:18:33 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:38.340 20:18:33 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:19:38.340 ************************************ 00:19:38.340 START TEST raid_function_test_concat 00:19:38.340 ************************************ 00:19:38.340 20:18:33 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1125 -- # raid_function_test concat 00:19:38.340 20:18:33 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@64 -- # local raid_level=concat 00:19:38.340 20:18:33 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:19:38.340 20:18:33 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:19:38.340 Process raid pid: 60734 00:19:38.340 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:38.340 20:18:33 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@69 -- # raid_pid=60734 00:19:38.340 20:18:33 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 60734' 00:19:38.340 20:18:33 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@71 -- # waitforlisten 60734 00:19:38.340 20:18:33 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:19:38.340 20:18:33 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@831 -- # '[' -z 60734 ']' 00:19:38.340 20:18:33 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:38.340 20:18:33 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:38.340 20:18:33 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:38.340 20:18:33 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:38.340 20:18:33 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:19:38.340 [2024-10-01 20:18:33.489561] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:19:38.340 [2024-10-01 20:18:33.490014] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:38.600 [2024-10-01 20:18:33.664226] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:38.858 [2024-10-01 20:18:33.904803] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:19:39.117 [2024-10-01 20:18:34.112380] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:39.117 [2024-10-01 20:18:34.112865] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:39.375 20:18:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:39.375 20:18:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@864 -- # return 0 00:19:39.375 20:18:34 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:19:39.375 20:18:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:39.375 20:18:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:19:39.375 Base_1 00:19:39.375 20:18:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:39.375 20:18:34 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:19:39.375 20:18:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:39.375 20:18:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:19:39.634 Base_2 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''Base_1 Base_2'\''' -n raid 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:19:39.634 [2024-10-01 20:18:34.647817] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:19:39.634 [2024-10-01 20:18:34.650432] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:19:39.634 [2024-10-01 20:18:34.650683] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:19:39.634 [2024-10-01 20:18:34.650743] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:19:39.634 [2024-10-01 20:18:34.651108] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:19:39.634 [2024-10-01 20:18:34.651307] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:19:39.634 [2024-10-01 20:18:34.651324] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000007780 00:19:39.634 [2024-10-01 20:18:34.651520] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@12 -- # local i 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:19:39.634 20:18:34 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:19:39.892 [2024-10-01 20:18:34.996110] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:19:39.892 /dev/nbd0 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@869 -- # local i 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@873 -- # break 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:19:39.892 1+0 records in 00:19:39.892 1+0 records out 00:19:39.892 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000540265 s, 7.6 MB/s 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@886 -- # size=4096 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@889 -- # return 0 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:19:39.892 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:19:40.151 { 00:19:40.151 "nbd_device": "/dev/nbd0", 00:19:40.151 "bdev_name": "raid" 00:19:40.151 } 00:19:40.151 ]' 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[ 00:19:40.151 { 00:19:40.151 "nbd_device": "/dev/nbd0", 00:19:40.151 "bdev_name": "raid" 00:19:40.151 } 00:19:40.151 ]' 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=1 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 1 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # count=1 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@19 -- # local blksize 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # blksize=512 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:19:40.151 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:19:40.409 4096+0 records in 00:19:40.409 4096+0 records out 00:19:40.409 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0264574 s, 79.3 MB/s 00:19:40.409 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:19:40.667 4096+0 records in 00:19:40.667 4096+0 records out 00:19:40.667 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.340493 s, 6.2 MB/s 00:19:40.667 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:19:40.667 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:19:40.667 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:19:40.667 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:19:40.667 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:19:40.667 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:19:40.667 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:19:40.667 128+0 records in 00:19:40.667 128+0 records out 00:19:40.667 65536 bytes (66 kB, 64 KiB) copied, 0.00067228 s, 97.5 MB/s 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:19:40.668 2035+0 records in 00:19:40.668 2035+0 records out 00:19:40.668 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.00882086 s, 118 MB/s 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:19:40.668 456+0 records in 00:19:40.668 456+0 records out 00:19:40.668 233472 bytes (233 kB, 228 KiB) copied, 0.00290515 s, 80.4 MB/s 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@52 -- # return 0 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@51 -- # local i 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:19:40.668 20:18:35 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:19:41.234 [2024-10-01 20:18:36.186496] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:41.234 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:19:41.234 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:19:41.234 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:19:41.234 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:19:41.234 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:19:41.234 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:19:41.234 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@41 -- # break 00:19:41.234 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@45 -- # return 0 00:19:41.234 20:18:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:19:41.234 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:19:41.234 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo '' 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # true 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=0 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 0 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # count=0 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@97 -- # killprocess 60734 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@950 -- # '[' -z 60734 ']' 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@954 -- # kill -0 60734 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@955 -- # uname 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 60734 00:19:41.493 killing process with pid 60734 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@968 -- # echo 'killing process with pid 60734' 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@969 -- # kill 60734 00:19:41.493 20:18:36 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@974 -- # wait 60734 00:19:41.493 [2024-10-01 20:18:36.594536] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:19:41.493 [2024-10-01 20:18:36.594675] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:19:41.493 [2024-10-01 20:18:36.594745] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:19:41.493 [2024-10-01 20:18:36.594766] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid, state offline 00:19:41.752 [2024-10-01 20:18:36.780594] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:19:43.656 20:18:38 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@99 -- # return 0 00:19:43.656 00:19:43.656 real 0m5.267s 00:19:43.656 user 0m6.095s 00:19:43.656 sys 0m1.219s 00:19:43.656 ************************************ 00:19:43.656 END TEST raid_function_test_concat 00:19:43.656 ************************************ 00:19:43.656 20:18:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:43.656 20:18:38 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:19:43.656 20:18:38 bdev_raid -- bdev/bdev_raid.sh@963 -- # run_test raid0_resize_test raid_resize_test 0 00:19:43.656 20:18:38 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:19:43.656 20:18:38 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:43.656 20:18:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:19:43.656 ************************************ 00:19:43.656 START TEST raid0_resize_test 00:19:43.656 ************************************ 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1125 -- # raid_resize_test 0 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=0 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:19:43.656 Process raid pid: 60874 00:19:43.656 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=60874 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 60874' 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 60874 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@831 -- # '[' -z 60874 ']' 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:43.656 20:18:38 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:43.656 [2024-10-01 20:18:38.813898] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:19:43.656 [2024-10-01 20:18:38.814327] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:43.917 [2024-10-01 20:18:38.994476] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:44.176 [2024-10-01 20:18:39.233213] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:19:44.434 [2024-10-01 20:18:39.435985] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:44.434 [2024-10-01 20:18:39.436386] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@864 -- # return 0 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:44.745 Base_1 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:44.745 Base_2 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 0 -eq 0 ']' 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@350 -- # rpc_cmd bdev_raid_create -z 64 -r 0 -b ''\''Base_1 Base_2'\''' -n Raid 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:44.745 [2024-10-01 20:18:39.873291] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:19:44.745 [2024-10-01 20:18:39.875871] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:19:44.745 [2024-10-01 20:18:39.876112] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:19:44.745 [2024-10-01 20:18:39.876142] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:19:44.745 [2024-10-01 20:18:39.876456] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:19:44.745 [2024-10-01 20:18:39.876619] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:19:44.745 [2024-10-01 20:18:39.876638] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:19:44.745 [2024-10-01 20:18:39.876847] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:44.745 [2024-10-01 20:18:39.881249] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:19:44.745 [2024-10-01 20:18:39.881286] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:19:44.745 true 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:19:44.745 [2024-10-01 20:18:39.893447] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=131072 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=64 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 0 -eq 0 ']' 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@362 -- # expected_size=64 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 64 '!=' 64 ']' 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:44.745 [2024-10-01 20:18:39.941312] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:19:44.745 [2024-10-01 20:18:39.941345] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:19:44.745 [2024-10-01 20:18:39.941407] bdev_raid.c:2340:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 131072 to 262144 00:19:44.745 true 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:19:44.745 [2024-10-01 20:18:39.953461] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:19:44.745 20:18:39 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:45.008 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=262144 00:19:45.008 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=128 00:19:45.008 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 0 -eq 0 ']' 00:19:45.008 20:18:39 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@378 -- # expected_size=128 00:19:45.008 20:18:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 128 '!=' 128 ']' 00:19:45.008 20:18:40 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 60874 00:19:45.008 20:18:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@950 -- # '[' -z 60874 ']' 00:19:45.008 20:18:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@954 -- # kill -0 60874 00:19:45.008 20:18:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@955 -- # uname 00:19:45.008 20:18:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:45.008 20:18:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 60874 00:19:45.008 killing process with pid 60874 00:19:45.008 20:18:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:19:45.008 20:18:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:19:45.008 20:18:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 60874' 00:19:45.008 20:18:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@969 -- # kill 60874 00:19:45.008 [2024-10-01 20:18:40.036604] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:19:45.008 20:18:40 bdev_raid.raid0_resize_test -- common/autotest_common.sh@974 -- # wait 60874 00:19:45.008 [2024-10-01 20:18:40.036745] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:19:45.008 [2024-10-01 20:18:40.036818] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:19:45.008 [2024-10-01 20:18:40.036834] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:19:45.008 [2024-10-01 20:18:40.052626] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:19:46.909 20:18:41 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:19:46.909 00:19:46.909 real 0m3.139s 00:19:46.909 user 0m3.229s 00:19:46.909 sys 0m0.542s 00:19:46.909 20:18:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:46.909 20:18:41 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:46.909 ************************************ 00:19:46.909 END TEST raid0_resize_test 00:19:46.909 ************************************ 00:19:46.909 20:18:41 bdev_raid -- bdev/bdev_raid.sh@964 -- # run_test raid1_resize_test raid_resize_test 1 00:19:46.909 20:18:41 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:19:46.909 20:18:41 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:46.909 20:18:41 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:19:46.909 ************************************ 00:19:46.909 START TEST raid1_resize_test 00:19:46.909 ************************************ 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1125 -- # raid_resize_test 1 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=1 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:19:46.909 Process raid pid: 60947 00:19:46.909 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=60947 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 60947' 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 60947 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- common/autotest_common.sh@831 -- # '[' -z 60947 ']' 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:46.909 20:18:41 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:46.909 [2024-10-01 20:18:42.001932] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:19:46.909 [2024-10-01 20:18:42.002438] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:47.167 [2024-10-01 20:18:42.177759] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:47.168 [2024-10-01 20:18:42.415628] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:19:47.425 [2024-10-01 20:18:42.619120] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:47.425 [2024-10-01 20:18:42.619454] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:47.994 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@864 -- # return 0 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:47.995 Base_1 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:47.995 Base_2 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 1 -eq 0 ']' 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@352 -- # rpc_cmd bdev_raid_create -r 1 -b ''\''Base_1 Base_2'\''' -n Raid 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:47.995 [2024-10-01 20:18:43.057005] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:19:47.995 [2024-10-01 20:18:43.059751] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:19:47.995 [2024-10-01 20:18:43.059976] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:19:47.995 [2024-10-01 20:18:43.060107] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:19:47.995 [2024-10-01 20:18:43.060479] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ba0 00:19:47.995 [2024-10-01 20:18:43.060783] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:19:47.995 [2024-10-01 20:18:43.060917] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000007780 00:19:47.995 [2024-10-01 20:18:43.061297] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:47.995 [2024-10-01 20:18:43.065248] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:19:47.995 [2024-10-01 20:18:43.065404] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:19:47.995 true 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:47.995 [2024-10-01 20:18:43.077414] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=65536 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=32 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 1 -eq 0 ']' 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@364 -- # expected_size=32 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 32 '!=' 32 ']' 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:47.995 [2024-10-01 20:18:43.125283] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:19:47.995 [2024-10-01 20:18:43.125330] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:19:47.995 [2024-10-01 20:18:43.125404] bdev_raid.c:2340:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 65536 to 131072 00:19:47.995 true 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:47.995 [2024-10-01 20:18:43.137451] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=131072 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=64 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 1 -eq 0 ']' 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@380 -- # expected_size=64 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 64 '!=' 64 ']' 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 60947 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@950 -- # '[' -z 60947 ']' 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@954 -- # kill -0 60947 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@955 -- # uname 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 60947 00:19:47.995 killing process with pid 60947 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 60947' 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@969 -- # kill 60947 00:19:47.995 [2024-10-01 20:18:43.214801] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:19:47.995 20:18:43 bdev_raid.raid1_resize_test -- common/autotest_common.sh@974 -- # wait 60947 00:19:47.995 [2024-10-01 20:18:43.214927] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:19:47.995 [2024-10-01 20:18:43.215584] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:19:47.995 [2024-10-01 20:18:43.215759] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Raid, state offline 00:19:47.995 [2024-10-01 20:18:43.230781] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:19:49.897 ************************************ 00:19:49.897 END TEST raid1_resize_test 00:19:49.897 ************************************ 00:19:49.897 20:18:45 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:19:49.897 00:19:49.897 real 0m3.116s 00:19:49.897 user 0m3.211s 00:19:49.897 sys 0m0.500s 00:19:49.897 20:18:45 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:49.897 20:18:45 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:19:49.897 20:18:45 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:19:49.897 20:18:45 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:19:49.897 20:18:45 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 2 false 00:19:49.897 20:18:45 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:19:49.897 20:18:45 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:49.897 20:18:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:19:49.897 ************************************ 00:19:49.897 START TEST raid_state_function_test 00:19:49.897 ************************************ 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 2 false 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:19:49.897 Process raid pid: 61015 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=61015 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 61015' 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 61015 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 61015 ']' 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:49.897 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:49.897 20:18:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:50.155 [2024-10-01 20:18:45.227958] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:19:50.155 [2024-10-01 20:18:45.228501] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:50.413 [2024-10-01 20:18:45.408770] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:50.413 [2024-10-01 20:18:45.646747] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:19:50.671 [2024-10-01 20:18:45.854041] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:50.671 [2024-10-01 20:18:45.854106] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:51.238 [2024-10-01 20:18:46.268285] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:19:51.238 [2024-10-01 20:18:46.268368] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:19:51.238 [2024-10-01 20:18:46.268386] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:19:51.238 [2024-10-01 20:18:46.268404] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:19:51.238 "name": "Existed_Raid", 00:19:51.238 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:51.238 "strip_size_kb": 64, 00:19:51.238 "state": "configuring", 00:19:51.238 "raid_level": "raid0", 00:19:51.238 "superblock": false, 00:19:51.238 "num_base_bdevs": 2, 00:19:51.238 "num_base_bdevs_discovered": 0, 00:19:51.238 "num_base_bdevs_operational": 2, 00:19:51.238 "base_bdevs_list": [ 00:19:51.238 { 00:19:51.238 "name": "BaseBdev1", 00:19:51.238 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:51.238 "is_configured": false, 00:19:51.238 "data_offset": 0, 00:19:51.238 "data_size": 0 00:19:51.238 }, 00:19:51.238 { 00:19:51.238 "name": "BaseBdev2", 00:19:51.238 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:51.238 "is_configured": false, 00:19:51.238 "data_offset": 0, 00:19:51.238 "data_size": 0 00:19:51.238 } 00:19:51.238 ] 00:19:51.238 }' 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:19:51.238 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:51.806 [2024-10-01 20:18:46.836353] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:19:51.806 [2024-10-01 20:18:46.836401] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:51.806 [2024-10-01 20:18:46.844346] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:19:51.806 [2024-10-01 20:18:46.844404] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:19:51.806 [2024-10-01 20:18:46.844421] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:19:51.806 [2024-10-01 20:18:46.844441] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:51.806 [2024-10-01 20:18:46.889955] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:19:51.806 BaseBdev1 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:51.806 [ 00:19:51.806 { 00:19:51.806 "name": "BaseBdev1", 00:19:51.806 "aliases": [ 00:19:51.806 "c227fbe8-e8d4-40db-95a5-f6c2d1ef7f38" 00:19:51.806 ], 00:19:51.806 "product_name": "Malloc disk", 00:19:51.806 "block_size": 512, 00:19:51.806 "num_blocks": 65536, 00:19:51.806 "uuid": "c227fbe8-e8d4-40db-95a5-f6c2d1ef7f38", 00:19:51.806 "assigned_rate_limits": { 00:19:51.806 "rw_ios_per_sec": 0, 00:19:51.806 "rw_mbytes_per_sec": 0, 00:19:51.806 "r_mbytes_per_sec": 0, 00:19:51.806 "w_mbytes_per_sec": 0 00:19:51.806 }, 00:19:51.806 "claimed": true, 00:19:51.806 "claim_type": "exclusive_write", 00:19:51.806 "zoned": false, 00:19:51.806 "supported_io_types": { 00:19:51.806 "read": true, 00:19:51.806 "write": true, 00:19:51.806 "unmap": true, 00:19:51.806 "flush": true, 00:19:51.806 "reset": true, 00:19:51.806 "nvme_admin": false, 00:19:51.806 "nvme_io": false, 00:19:51.806 "nvme_io_md": false, 00:19:51.806 "write_zeroes": true, 00:19:51.806 "zcopy": true, 00:19:51.806 "get_zone_info": false, 00:19:51.806 "zone_management": false, 00:19:51.806 "zone_append": false, 00:19:51.806 "compare": false, 00:19:51.806 "compare_and_write": false, 00:19:51.806 "abort": true, 00:19:51.806 "seek_hole": false, 00:19:51.806 "seek_data": false, 00:19:51.806 "copy": true, 00:19:51.806 "nvme_iov_md": false 00:19:51.806 }, 00:19:51.806 "memory_domains": [ 00:19:51.806 { 00:19:51.806 "dma_device_id": "system", 00:19:51.806 "dma_device_type": 1 00:19:51.806 }, 00:19:51.806 { 00:19:51.806 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:51.806 "dma_device_type": 2 00:19:51.806 } 00:19:51.806 ], 00:19:51.806 "driver_specific": {} 00:19:51.806 } 00:19:51.806 ] 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:19:51.806 "name": "Existed_Raid", 00:19:51.806 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:51.806 "strip_size_kb": 64, 00:19:51.806 "state": "configuring", 00:19:51.806 "raid_level": "raid0", 00:19:51.806 "superblock": false, 00:19:51.806 "num_base_bdevs": 2, 00:19:51.806 "num_base_bdevs_discovered": 1, 00:19:51.806 "num_base_bdevs_operational": 2, 00:19:51.806 "base_bdevs_list": [ 00:19:51.806 { 00:19:51.806 "name": "BaseBdev1", 00:19:51.806 "uuid": "c227fbe8-e8d4-40db-95a5-f6c2d1ef7f38", 00:19:51.806 "is_configured": true, 00:19:51.806 "data_offset": 0, 00:19:51.806 "data_size": 65536 00:19:51.806 }, 00:19:51.806 { 00:19:51.806 "name": "BaseBdev2", 00:19:51.806 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:51.806 "is_configured": false, 00:19:51.806 "data_offset": 0, 00:19:51.806 "data_size": 0 00:19:51.806 } 00:19:51.806 ] 00:19:51.806 }' 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:19:51.806 20:18:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:52.373 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:19:52.373 20:18:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:52.373 20:18:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:52.374 [2024-10-01 20:18:47.458218] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:19:52.374 [2024-10-01 20:18:47.458288] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:52.374 [2024-10-01 20:18:47.466208] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:19:52.374 [2024-10-01 20:18:47.468940] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:19:52.374 [2024-10-01 20:18:47.469164] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:19:52.374 "name": "Existed_Raid", 00:19:52.374 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:52.374 "strip_size_kb": 64, 00:19:52.374 "state": "configuring", 00:19:52.374 "raid_level": "raid0", 00:19:52.374 "superblock": false, 00:19:52.374 "num_base_bdevs": 2, 00:19:52.374 "num_base_bdevs_discovered": 1, 00:19:52.374 "num_base_bdevs_operational": 2, 00:19:52.374 "base_bdevs_list": [ 00:19:52.374 { 00:19:52.374 "name": "BaseBdev1", 00:19:52.374 "uuid": "c227fbe8-e8d4-40db-95a5-f6c2d1ef7f38", 00:19:52.374 "is_configured": true, 00:19:52.374 "data_offset": 0, 00:19:52.374 "data_size": 65536 00:19:52.374 }, 00:19:52.374 { 00:19:52.374 "name": "BaseBdev2", 00:19:52.374 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:52.374 "is_configured": false, 00:19:52.374 "data_offset": 0, 00:19:52.374 "data_size": 0 00:19:52.374 } 00:19:52.374 ] 00:19:52.374 }' 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:19:52.374 20:18:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:52.942 20:18:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:19:52.942 20:18:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:52.942 20:18:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:52.942 [2024-10-01 20:18:48.038955] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:19:52.942 [2024-10-01 20:18:48.039273] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:19:52.942 [2024-10-01 20:18:48.039415] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:19:52.942 [2024-10-01 20:18:48.039984] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:19:52.942 [2024-10-01 20:18:48.040335] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:19:52.942 [2024-10-01 20:18:48.040368] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:19:52.942 [2024-10-01 20:18:48.040786] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:52.942 BaseBdev2 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:52.942 [ 00:19:52.942 { 00:19:52.942 "name": "BaseBdev2", 00:19:52.942 "aliases": [ 00:19:52.942 "3298fe86-24fd-4fab-b864-b455856af58f" 00:19:52.942 ], 00:19:52.942 "product_name": "Malloc disk", 00:19:52.942 "block_size": 512, 00:19:52.942 "num_blocks": 65536, 00:19:52.942 "uuid": "3298fe86-24fd-4fab-b864-b455856af58f", 00:19:52.942 "assigned_rate_limits": { 00:19:52.942 "rw_ios_per_sec": 0, 00:19:52.942 "rw_mbytes_per_sec": 0, 00:19:52.942 "r_mbytes_per_sec": 0, 00:19:52.942 "w_mbytes_per_sec": 0 00:19:52.942 }, 00:19:52.942 "claimed": true, 00:19:52.942 "claim_type": "exclusive_write", 00:19:52.942 "zoned": false, 00:19:52.942 "supported_io_types": { 00:19:52.942 "read": true, 00:19:52.942 "write": true, 00:19:52.942 "unmap": true, 00:19:52.942 "flush": true, 00:19:52.942 "reset": true, 00:19:52.942 "nvme_admin": false, 00:19:52.942 "nvme_io": false, 00:19:52.942 "nvme_io_md": false, 00:19:52.942 "write_zeroes": true, 00:19:52.942 "zcopy": true, 00:19:52.942 "get_zone_info": false, 00:19:52.942 "zone_management": false, 00:19:52.942 "zone_append": false, 00:19:52.942 "compare": false, 00:19:52.942 "compare_and_write": false, 00:19:52.942 "abort": true, 00:19:52.942 "seek_hole": false, 00:19:52.942 "seek_data": false, 00:19:52.942 "copy": true, 00:19:52.942 "nvme_iov_md": false 00:19:52.942 }, 00:19:52.942 "memory_domains": [ 00:19:52.942 { 00:19:52.942 "dma_device_id": "system", 00:19:52.942 "dma_device_type": 1 00:19:52.942 }, 00:19:52.942 { 00:19:52.942 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:52.942 "dma_device_type": 2 00:19:52.942 } 00:19:52.942 ], 00:19:52.942 "driver_specific": {} 00:19:52.942 } 00:19:52.942 ] 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:19:52.942 "name": "Existed_Raid", 00:19:52.942 "uuid": "a6216456-094b-49e4-8478-01ba5c0eccbf", 00:19:52.942 "strip_size_kb": 64, 00:19:52.942 "state": "online", 00:19:52.942 "raid_level": "raid0", 00:19:52.942 "superblock": false, 00:19:52.942 "num_base_bdevs": 2, 00:19:52.942 "num_base_bdevs_discovered": 2, 00:19:52.942 "num_base_bdevs_operational": 2, 00:19:52.942 "base_bdevs_list": [ 00:19:52.942 { 00:19:52.942 "name": "BaseBdev1", 00:19:52.942 "uuid": "c227fbe8-e8d4-40db-95a5-f6c2d1ef7f38", 00:19:52.942 "is_configured": true, 00:19:52.942 "data_offset": 0, 00:19:52.942 "data_size": 65536 00:19:52.942 }, 00:19:52.942 { 00:19:52.942 "name": "BaseBdev2", 00:19:52.942 "uuid": "3298fe86-24fd-4fab-b864-b455856af58f", 00:19:52.942 "is_configured": true, 00:19:52.942 "data_offset": 0, 00:19:52.942 "data_size": 65536 00:19:52.942 } 00:19:52.942 ] 00:19:52.942 }' 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:19:52.942 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:53.510 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:19:53.510 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:19:53.510 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:19:53.510 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:53.511 [2024-10-01 20:18:48.599506] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:19:53.511 "name": "Existed_Raid", 00:19:53.511 "aliases": [ 00:19:53.511 "a6216456-094b-49e4-8478-01ba5c0eccbf" 00:19:53.511 ], 00:19:53.511 "product_name": "Raid Volume", 00:19:53.511 "block_size": 512, 00:19:53.511 "num_blocks": 131072, 00:19:53.511 "uuid": "a6216456-094b-49e4-8478-01ba5c0eccbf", 00:19:53.511 "assigned_rate_limits": { 00:19:53.511 "rw_ios_per_sec": 0, 00:19:53.511 "rw_mbytes_per_sec": 0, 00:19:53.511 "r_mbytes_per_sec": 0, 00:19:53.511 "w_mbytes_per_sec": 0 00:19:53.511 }, 00:19:53.511 "claimed": false, 00:19:53.511 "zoned": false, 00:19:53.511 "supported_io_types": { 00:19:53.511 "read": true, 00:19:53.511 "write": true, 00:19:53.511 "unmap": true, 00:19:53.511 "flush": true, 00:19:53.511 "reset": true, 00:19:53.511 "nvme_admin": false, 00:19:53.511 "nvme_io": false, 00:19:53.511 "nvme_io_md": false, 00:19:53.511 "write_zeroes": true, 00:19:53.511 "zcopy": false, 00:19:53.511 "get_zone_info": false, 00:19:53.511 "zone_management": false, 00:19:53.511 "zone_append": false, 00:19:53.511 "compare": false, 00:19:53.511 "compare_and_write": false, 00:19:53.511 "abort": false, 00:19:53.511 "seek_hole": false, 00:19:53.511 "seek_data": false, 00:19:53.511 "copy": false, 00:19:53.511 "nvme_iov_md": false 00:19:53.511 }, 00:19:53.511 "memory_domains": [ 00:19:53.511 { 00:19:53.511 "dma_device_id": "system", 00:19:53.511 "dma_device_type": 1 00:19:53.511 }, 00:19:53.511 { 00:19:53.511 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:53.511 "dma_device_type": 2 00:19:53.511 }, 00:19:53.511 { 00:19:53.511 "dma_device_id": "system", 00:19:53.511 "dma_device_type": 1 00:19:53.511 }, 00:19:53.511 { 00:19:53.511 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:53.511 "dma_device_type": 2 00:19:53.511 } 00:19:53.511 ], 00:19:53.511 "driver_specific": { 00:19:53.511 "raid": { 00:19:53.511 "uuid": "a6216456-094b-49e4-8478-01ba5c0eccbf", 00:19:53.511 "strip_size_kb": 64, 00:19:53.511 "state": "online", 00:19:53.511 "raid_level": "raid0", 00:19:53.511 "superblock": false, 00:19:53.511 "num_base_bdevs": 2, 00:19:53.511 "num_base_bdevs_discovered": 2, 00:19:53.511 "num_base_bdevs_operational": 2, 00:19:53.511 "base_bdevs_list": [ 00:19:53.511 { 00:19:53.511 "name": "BaseBdev1", 00:19:53.511 "uuid": "c227fbe8-e8d4-40db-95a5-f6c2d1ef7f38", 00:19:53.511 "is_configured": true, 00:19:53.511 "data_offset": 0, 00:19:53.511 "data_size": 65536 00:19:53.511 }, 00:19:53.511 { 00:19:53.511 "name": "BaseBdev2", 00:19:53.511 "uuid": "3298fe86-24fd-4fab-b864-b455856af58f", 00:19:53.511 "is_configured": true, 00:19:53.511 "data_offset": 0, 00:19:53.511 "data_size": 65536 00:19:53.511 } 00:19:53.511 ] 00:19:53.511 } 00:19:53.511 } 00:19:53.511 }' 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:19:53.511 BaseBdev2' 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:53.511 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:53.833 [2024-10-01 20:18:48.863316] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:19:53.833 [2024-10-01 20:18:48.863521] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:19:53.833 [2024-10-01 20:18:48.863621] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:19:53.833 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:19:53.834 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:53.834 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:53.834 20:18:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:53.834 20:18:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:53.834 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:19:53.834 "name": "Existed_Raid", 00:19:53.834 "uuid": "a6216456-094b-49e4-8478-01ba5c0eccbf", 00:19:53.834 "strip_size_kb": 64, 00:19:53.834 "state": "offline", 00:19:53.834 "raid_level": "raid0", 00:19:53.834 "superblock": false, 00:19:53.834 "num_base_bdevs": 2, 00:19:53.834 "num_base_bdevs_discovered": 1, 00:19:53.834 "num_base_bdevs_operational": 1, 00:19:53.834 "base_bdevs_list": [ 00:19:53.834 { 00:19:53.834 "name": null, 00:19:53.834 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:53.834 "is_configured": false, 00:19:53.834 "data_offset": 0, 00:19:53.834 "data_size": 65536 00:19:53.834 }, 00:19:53.834 { 00:19:53.834 "name": "BaseBdev2", 00:19:53.834 "uuid": "3298fe86-24fd-4fab-b864-b455856af58f", 00:19:53.834 "is_configured": true, 00:19:53.834 "data_offset": 0, 00:19:53.834 "data_size": 65536 00:19:53.834 } 00:19:53.834 ] 00:19:53.834 }' 00:19:53.834 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:19:53.834 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:54.401 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:19:54.401 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:19:54.401 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:19:54.401 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:19:54.401 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:54.401 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:54.401 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:54.401 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:19:54.401 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:19:54.401 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:19:54.401 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:54.401 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:54.401 [2024-10-01 20:18:49.565034] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:19:54.401 [2024-10-01 20:18:49.565257] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:19:54.660 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:54.660 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 61015 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 61015 ']' 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 61015 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 61015 00:19:54.661 killing process with pid 61015 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 61015' 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 61015 00:19:54.661 [2024-10-01 20:18:49.747004] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:19:54.661 20:18:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 61015 00:19:54.661 [2024-10-01 20:18:49.762146] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:19:56.562 00:19:56.562 real 0m6.447s 00:19:56.562 user 0m9.243s 00:19:56.562 sys 0m0.938s 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:19:56.562 ************************************ 00:19:56.562 END TEST raid_state_function_test 00:19:56.562 ************************************ 00:19:56.562 20:18:51 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 2 true 00:19:56.562 20:18:51 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:19:56.562 20:18:51 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:56.562 20:18:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:19:56.562 ************************************ 00:19:56.562 START TEST raid_state_function_test_sb 00:19:56.562 ************************************ 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 2 true 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:19:56.562 Process raid pid: 61279 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=61279 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 61279' 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 61279 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 61279 ']' 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:56.562 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:56.562 20:18:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:56.562 [2024-10-01 20:18:51.683069] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:19:56.562 [2024-10-01 20:18:51.683529] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:56.821 [2024-10-01 20:18:51.865200] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:57.080 [2024-10-01 20:18:52.101462] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:19:57.080 [2024-10-01 20:18:52.290566] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:57.080 [2024-10-01 20:18:52.290979] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:57.647 [2024-10-01 20:18:52.709601] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:19:57.647 [2024-10-01 20:18:52.709687] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:19:57.647 [2024-10-01 20:18:52.709705] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:19:57.647 [2024-10-01 20:18:52.709738] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:19:57.647 "name": "Existed_Raid", 00:19:57.647 "uuid": "f1031d8b-527e-4262-85d8-f7490dd3c9a5", 00:19:57.647 "strip_size_kb": 64, 00:19:57.647 "state": "configuring", 00:19:57.647 "raid_level": "raid0", 00:19:57.647 "superblock": true, 00:19:57.647 "num_base_bdevs": 2, 00:19:57.647 "num_base_bdevs_discovered": 0, 00:19:57.647 "num_base_bdevs_operational": 2, 00:19:57.647 "base_bdevs_list": [ 00:19:57.647 { 00:19:57.647 "name": "BaseBdev1", 00:19:57.647 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:57.647 "is_configured": false, 00:19:57.647 "data_offset": 0, 00:19:57.647 "data_size": 0 00:19:57.647 }, 00:19:57.647 { 00:19:57.647 "name": "BaseBdev2", 00:19:57.647 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:57.647 "is_configured": false, 00:19:57.647 "data_offset": 0, 00:19:57.647 "data_size": 0 00:19:57.647 } 00:19:57.647 ] 00:19:57.647 }' 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:19:57.647 20:18:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:58.215 [2024-10-01 20:18:53.229595] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:19:58.215 [2024-10-01 20:18:53.229641] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:58.215 [2024-10-01 20:18:53.237614] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:19:58.215 [2024-10-01 20:18:53.237683] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:19:58.215 [2024-10-01 20:18:53.237699] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:19:58.215 [2024-10-01 20:18:53.237736] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:58.215 [2024-10-01 20:18:53.280398] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:19:58.215 BaseBdev1 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:58.215 [ 00:19:58.215 { 00:19:58.215 "name": "BaseBdev1", 00:19:58.215 "aliases": [ 00:19:58.215 "914f2564-4d91-493d-a357-5be376e0753e" 00:19:58.215 ], 00:19:58.215 "product_name": "Malloc disk", 00:19:58.215 "block_size": 512, 00:19:58.215 "num_blocks": 65536, 00:19:58.215 "uuid": "914f2564-4d91-493d-a357-5be376e0753e", 00:19:58.215 "assigned_rate_limits": { 00:19:58.215 "rw_ios_per_sec": 0, 00:19:58.215 "rw_mbytes_per_sec": 0, 00:19:58.215 "r_mbytes_per_sec": 0, 00:19:58.215 "w_mbytes_per_sec": 0 00:19:58.215 }, 00:19:58.215 "claimed": true, 00:19:58.215 "claim_type": "exclusive_write", 00:19:58.215 "zoned": false, 00:19:58.215 "supported_io_types": { 00:19:58.215 "read": true, 00:19:58.215 "write": true, 00:19:58.215 "unmap": true, 00:19:58.215 "flush": true, 00:19:58.215 "reset": true, 00:19:58.215 "nvme_admin": false, 00:19:58.215 "nvme_io": false, 00:19:58.215 "nvme_io_md": false, 00:19:58.215 "write_zeroes": true, 00:19:58.215 "zcopy": true, 00:19:58.215 "get_zone_info": false, 00:19:58.215 "zone_management": false, 00:19:58.215 "zone_append": false, 00:19:58.215 "compare": false, 00:19:58.215 "compare_and_write": false, 00:19:58.215 "abort": true, 00:19:58.215 "seek_hole": false, 00:19:58.215 "seek_data": false, 00:19:58.215 "copy": true, 00:19:58.215 "nvme_iov_md": false 00:19:58.215 }, 00:19:58.215 "memory_domains": [ 00:19:58.215 { 00:19:58.215 "dma_device_id": "system", 00:19:58.215 "dma_device_type": 1 00:19:58.215 }, 00:19:58.215 { 00:19:58.215 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:58.215 "dma_device_type": 2 00:19:58.215 } 00:19:58.215 ], 00:19:58.215 "driver_specific": {} 00:19:58.215 } 00:19:58.215 ] 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:19:58.215 "name": "Existed_Raid", 00:19:58.215 "uuid": "40f8c837-e3bb-4924-bb52-3fac078ae149", 00:19:58.215 "strip_size_kb": 64, 00:19:58.215 "state": "configuring", 00:19:58.215 "raid_level": "raid0", 00:19:58.215 "superblock": true, 00:19:58.215 "num_base_bdevs": 2, 00:19:58.215 "num_base_bdevs_discovered": 1, 00:19:58.215 "num_base_bdevs_operational": 2, 00:19:58.215 "base_bdevs_list": [ 00:19:58.215 { 00:19:58.215 "name": "BaseBdev1", 00:19:58.215 "uuid": "914f2564-4d91-493d-a357-5be376e0753e", 00:19:58.215 "is_configured": true, 00:19:58.215 "data_offset": 2048, 00:19:58.215 "data_size": 63488 00:19:58.215 }, 00:19:58.215 { 00:19:58.215 "name": "BaseBdev2", 00:19:58.215 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:58.215 "is_configured": false, 00:19:58.215 "data_offset": 0, 00:19:58.215 "data_size": 0 00:19:58.215 } 00:19:58.215 ] 00:19:58.215 }' 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:19:58.215 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:58.783 [2024-10-01 20:18:53.852630] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:19:58.783 [2024-10-01 20:18:53.852906] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:58.783 [2024-10-01 20:18:53.860708] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:19:58.783 [2024-10-01 20:18:53.863388] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:19:58.783 [2024-10-01 20:18:53.863460] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:58.783 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:19:58.783 "name": "Existed_Raid", 00:19:58.783 "uuid": "19abbf6f-b7a5-4958-8de0-0d1658e27f95", 00:19:58.783 "strip_size_kb": 64, 00:19:58.783 "state": "configuring", 00:19:58.783 "raid_level": "raid0", 00:19:58.783 "superblock": true, 00:19:58.783 "num_base_bdevs": 2, 00:19:58.783 "num_base_bdevs_discovered": 1, 00:19:58.783 "num_base_bdevs_operational": 2, 00:19:58.783 "base_bdevs_list": [ 00:19:58.783 { 00:19:58.783 "name": "BaseBdev1", 00:19:58.783 "uuid": "914f2564-4d91-493d-a357-5be376e0753e", 00:19:58.784 "is_configured": true, 00:19:58.784 "data_offset": 2048, 00:19:58.784 "data_size": 63488 00:19:58.784 }, 00:19:58.784 { 00:19:58.784 "name": "BaseBdev2", 00:19:58.784 "uuid": "00000000-0000-0000-0000-000000000000", 00:19:58.784 "is_configured": false, 00:19:58.784 "data_offset": 0, 00:19:58.784 "data_size": 0 00:19:58.784 } 00:19:58.784 ] 00:19:58.784 }' 00:19:58.784 20:18:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:19:58.784 20:18:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:59.351 [2024-10-01 20:18:54.410088] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:19:59.351 [2024-10-01 20:18:54.410691] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:19:59.351 BaseBdev2 00:19:59.351 [2024-10-01 20:18:54.410886] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:19:59.351 [2024-10-01 20:18:54.411247] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:19:59.351 [2024-10-01 20:18:54.411475] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:19:59.351 [2024-10-01 20:18:54.411496] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:59.351 [2024-10-01 20:18:54.411665] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:59.351 [ 00:19:59.351 { 00:19:59.351 "name": "BaseBdev2", 00:19:59.351 "aliases": [ 00:19:59.351 "6488dea0-4985-45ad-874c-7d9e44236eaf" 00:19:59.351 ], 00:19:59.351 "product_name": "Malloc disk", 00:19:59.351 "block_size": 512, 00:19:59.351 "num_blocks": 65536, 00:19:59.351 "uuid": "6488dea0-4985-45ad-874c-7d9e44236eaf", 00:19:59.351 "assigned_rate_limits": { 00:19:59.351 "rw_ios_per_sec": 0, 00:19:59.351 "rw_mbytes_per_sec": 0, 00:19:59.351 "r_mbytes_per_sec": 0, 00:19:59.351 "w_mbytes_per_sec": 0 00:19:59.351 }, 00:19:59.351 "claimed": true, 00:19:59.351 "claim_type": "exclusive_write", 00:19:59.351 "zoned": false, 00:19:59.351 "supported_io_types": { 00:19:59.351 "read": true, 00:19:59.351 "write": true, 00:19:59.351 "unmap": true, 00:19:59.351 "flush": true, 00:19:59.351 "reset": true, 00:19:59.351 "nvme_admin": false, 00:19:59.351 "nvme_io": false, 00:19:59.351 "nvme_io_md": false, 00:19:59.351 "write_zeroes": true, 00:19:59.351 "zcopy": true, 00:19:59.351 "get_zone_info": false, 00:19:59.351 "zone_management": false, 00:19:59.351 "zone_append": false, 00:19:59.351 "compare": false, 00:19:59.351 "compare_and_write": false, 00:19:59.351 "abort": true, 00:19:59.351 "seek_hole": false, 00:19:59.351 "seek_data": false, 00:19:59.351 "copy": true, 00:19:59.351 "nvme_iov_md": false 00:19:59.351 }, 00:19:59.351 "memory_domains": [ 00:19:59.351 { 00:19:59.351 "dma_device_id": "system", 00:19:59.351 "dma_device_type": 1 00:19:59.351 }, 00:19:59.351 { 00:19:59.351 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:59.351 "dma_device_type": 2 00:19:59.351 } 00:19:59.351 ], 00:19:59.351 "driver_specific": {} 00:19:59.351 } 00:19:59.351 ] 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:59.351 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:19:59.351 "name": "Existed_Raid", 00:19:59.351 "uuid": "19abbf6f-b7a5-4958-8de0-0d1658e27f95", 00:19:59.351 "strip_size_kb": 64, 00:19:59.351 "state": "online", 00:19:59.351 "raid_level": "raid0", 00:19:59.351 "superblock": true, 00:19:59.351 "num_base_bdevs": 2, 00:19:59.351 "num_base_bdevs_discovered": 2, 00:19:59.351 "num_base_bdevs_operational": 2, 00:19:59.351 "base_bdevs_list": [ 00:19:59.351 { 00:19:59.352 "name": "BaseBdev1", 00:19:59.352 "uuid": "914f2564-4d91-493d-a357-5be376e0753e", 00:19:59.352 "is_configured": true, 00:19:59.352 "data_offset": 2048, 00:19:59.352 "data_size": 63488 00:19:59.352 }, 00:19:59.352 { 00:19:59.352 "name": "BaseBdev2", 00:19:59.352 "uuid": "6488dea0-4985-45ad-874c-7d9e44236eaf", 00:19:59.352 "is_configured": true, 00:19:59.352 "data_offset": 2048, 00:19:59.352 "data_size": 63488 00:19:59.352 } 00:19:59.352 ] 00:19:59.352 }' 00:19:59.352 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:19:59.352 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:59.919 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:19:59.919 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:19:59.919 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:19:59.919 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:19:59.919 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:19:59.919 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:19:59.919 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:19:59.919 20:18:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:19:59.919 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:59.919 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:59.919 [2024-10-01 20:18:54.974797] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:19:59.919 20:18:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:59.919 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:19:59.919 "name": "Existed_Raid", 00:19:59.919 "aliases": [ 00:19:59.919 "19abbf6f-b7a5-4958-8de0-0d1658e27f95" 00:19:59.919 ], 00:19:59.919 "product_name": "Raid Volume", 00:19:59.919 "block_size": 512, 00:19:59.919 "num_blocks": 126976, 00:19:59.919 "uuid": "19abbf6f-b7a5-4958-8de0-0d1658e27f95", 00:19:59.919 "assigned_rate_limits": { 00:19:59.919 "rw_ios_per_sec": 0, 00:19:59.919 "rw_mbytes_per_sec": 0, 00:19:59.919 "r_mbytes_per_sec": 0, 00:19:59.919 "w_mbytes_per_sec": 0 00:19:59.919 }, 00:19:59.919 "claimed": false, 00:19:59.919 "zoned": false, 00:19:59.919 "supported_io_types": { 00:19:59.919 "read": true, 00:19:59.919 "write": true, 00:19:59.919 "unmap": true, 00:19:59.919 "flush": true, 00:19:59.919 "reset": true, 00:19:59.919 "nvme_admin": false, 00:19:59.919 "nvme_io": false, 00:19:59.919 "nvme_io_md": false, 00:19:59.919 "write_zeroes": true, 00:19:59.919 "zcopy": false, 00:19:59.919 "get_zone_info": false, 00:19:59.919 "zone_management": false, 00:19:59.919 "zone_append": false, 00:19:59.919 "compare": false, 00:19:59.919 "compare_and_write": false, 00:19:59.919 "abort": false, 00:19:59.919 "seek_hole": false, 00:19:59.919 "seek_data": false, 00:19:59.919 "copy": false, 00:19:59.919 "nvme_iov_md": false 00:19:59.919 }, 00:19:59.919 "memory_domains": [ 00:19:59.919 { 00:19:59.919 "dma_device_id": "system", 00:19:59.919 "dma_device_type": 1 00:19:59.919 }, 00:19:59.919 { 00:19:59.919 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:59.919 "dma_device_type": 2 00:19:59.919 }, 00:19:59.919 { 00:19:59.919 "dma_device_id": "system", 00:19:59.919 "dma_device_type": 1 00:19:59.919 }, 00:19:59.919 { 00:19:59.919 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:19:59.919 "dma_device_type": 2 00:19:59.919 } 00:19:59.919 ], 00:19:59.919 "driver_specific": { 00:19:59.919 "raid": { 00:19:59.919 "uuid": "19abbf6f-b7a5-4958-8de0-0d1658e27f95", 00:19:59.919 "strip_size_kb": 64, 00:19:59.919 "state": "online", 00:19:59.919 "raid_level": "raid0", 00:19:59.919 "superblock": true, 00:19:59.919 "num_base_bdevs": 2, 00:19:59.919 "num_base_bdevs_discovered": 2, 00:19:59.919 "num_base_bdevs_operational": 2, 00:19:59.919 "base_bdevs_list": [ 00:19:59.919 { 00:19:59.919 "name": "BaseBdev1", 00:19:59.919 "uuid": "914f2564-4d91-493d-a357-5be376e0753e", 00:19:59.919 "is_configured": true, 00:19:59.919 "data_offset": 2048, 00:19:59.919 "data_size": 63488 00:19:59.919 }, 00:19:59.919 { 00:19:59.919 "name": "BaseBdev2", 00:19:59.919 "uuid": "6488dea0-4985-45ad-874c-7d9e44236eaf", 00:19:59.919 "is_configured": true, 00:19:59.919 "data_offset": 2048, 00:19:59.919 "data_size": 63488 00:19:59.919 } 00:19:59.919 ] 00:19:59.919 } 00:19:59.919 } 00:19:59.919 }' 00:19:59.919 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:19:59.919 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:19:59.919 BaseBdev2' 00:19:59.919 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:19:59.919 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:19:59.919 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:19:59.919 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:19:59.919 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:59.919 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:19:59.919 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:19:59.919 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:00.178 [2024-10-01 20:18:55.242514] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:20:00.178 [2024-10-01 20:18:55.242557] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:20:00.178 [2024-10-01 20:18:55.242627] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:00.178 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:00.179 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:00.179 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:00.179 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:00.179 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:00.179 "name": "Existed_Raid", 00:20:00.179 "uuid": "19abbf6f-b7a5-4958-8de0-0d1658e27f95", 00:20:00.179 "strip_size_kb": 64, 00:20:00.179 "state": "offline", 00:20:00.179 "raid_level": "raid0", 00:20:00.179 "superblock": true, 00:20:00.179 "num_base_bdevs": 2, 00:20:00.179 "num_base_bdevs_discovered": 1, 00:20:00.179 "num_base_bdevs_operational": 1, 00:20:00.179 "base_bdevs_list": [ 00:20:00.179 { 00:20:00.179 "name": null, 00:20:00.179 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:00.179 "is_configured": false, 00:20:00.179 "data_offset": 0, 00:20:00.179 "data_size": 63488 00:20:00.179 }, 00:20:00.179 { 00:20:00.179 "name": "BaseBdev2", 00:20:00.179 "uuid": "6488dea0-4985-45ad-874c-7d9e44236eaf", 00:20:00.179 "is_configured": true, 00:20:00.179 "data_offset": 2048, 00:20:00.179 "data_size": 63488 00:20:00.179 } 00:20:00.179 ] 00:20:00.179 }' 00:20:00.179 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:00.179 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:00.746 [2024-10-01 20:18:55.905896] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:20:00.746 [2024-10-01 20:18:55.905966] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:00.746 20:18:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:01.004 20:18:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:01.004 20:18:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:20:01.004 20:18:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:20:01.004 20:18:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:20:01.004 20:18:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 61279 00:20:01.004 20:18:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 61279 ']' 00:20:01.004 20:18:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 61279 00:20:01.004 20:18:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:20:01.004 20:18:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:20:01.004 20:18:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 61279 00:20:01.004 killing process with pid 61279 00:20:01.004 20:18:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:20:01.004 20:18:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:20:01.004 20:18:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 61279' 00:20:01.004 20:18:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 61279 00:20:01.004 [2024-10-01 20:18:56.085667] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:20:01.004 20:18:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 61279 00:20:01.004 [2024-10-01 20:18:56.100938] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:20:02.907 ************************************ 00:20:02.907 END TEST raid_state_function_test_sb 00:20:02.907 ************************************ 00:20:02.907 20:18:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:20:02.907 00:20:02.907 real 0m6.342s 00:20:02.907 user 0m9.061s 00:20:02.907 sys 0m0.922s 00:20:02.907 20:18:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:20:02.907 20:18:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:02.907 20:18:57 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 2 00:20:02.907 20:18:57 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:20:02.907 20:18:57 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:20:02.907 20:18:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:20:02.907 ************************************ 00:20:02.907 START TEST raid_superblock_test 00:20:02.907 ************************************ 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid0 2 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=61542 00:20:02.907 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 61542 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 61542 ']' 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:02.907 20:18:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:02.907 [2024-10-01 20:18:58.079601] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:20:02.907 [2024-10-01 20:18:58.079841] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid61542 ] 00:20:03.166 [2024-10-01 20:18:58.257091] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:03.426 [2024-10-01 20:18:58.507922] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:20:03.686 [2024-10-01 20:18:58.712353] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:03.686 [2024-10-01 20:18:58.712450] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:03.945 malloc1 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:03.945 [2024-10-01 20:18:59.165592] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:20:03.945 [2024-10-01 20:18:59.165869] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:03.945 [2024-10-01 20:18:59.165956] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:20:03.945 [2024-10-01 20:18:59.166205] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:03.945 [2024-10-01 20:18:59.169280] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:03.945 [2024-10-01 20:18:59.169484] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:20:03.945 pt1 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:03.945 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:04.204 malloc2 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:04.204 [2024-10-01 20:18:59.220218] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:20:04.204 [2024-10-01 20:18:59.220468] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:04.204 [2024-10-01 20:18:59.220551] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:20:04.204 [2024-10-01 20:18:59.220671] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:04.204 [2024-10-01 20:18:59.223619] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:04.204 [2024-10-01 20:18:59.223667] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:20:04.204 pt2 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:04.204 [2024-10-01 20:18:59.232348] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:20:04.204 [2024-10-01 20:18:59.235067] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:20:04.204 [2024-10-01 20:18:59.235524] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:20:04.204 [2024-10-01 20:18:59.235551] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:20:04.204 [2024-10-01 20:18:59.235920] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:20:04.204 [2024-10-01 20:18:59.236179] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:20:04.204 [2024-10-01 20:18:59.236199] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:20:04.204 [2024-10-01 20:18:59.236481] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:04.204 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:04.205 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:04.205 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:04.205 "name": "raid_bdev1", 00:20:04.205 "uuid": "2cfc7c6d-dd67-42c7-ad97-88ac1dcdf9e3", 00:20:04.205 "strip_size_kb": 64, 00:20:04.205 "state": "online", 00:20:04.205 "raid_level": "raid0", 00:20:04.205 "superblock": true, 00:20:04.205 "num_base_bdevs": 2, 00:20:04.205 "num_base_bdevs_discovered": 2, 00:20:04.205 "num_base_bdevs_operational": 2, 00:20:04.205 "base_bdevs_list": [ 00:20:04.205 { 00:20:04.205 "name": "pt1", 00:20:04.205 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:04.205 "is_configured": true, 00:20:04.205 "data_offset": 2048, 00:20:04.205 "data_size": 63488 00:20:04.205 }, 00:20:04.205 { 00:20:04.205 "name": "pt2", 00:20:04.205 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:04.205 "is_configured": true, 00:20:04.205 "data_offset": 2048, 00:20:04.205 "data_size": 63488 00:20:04.205 } 00:20:04.205 ] 00:20:04.205 }' 00:20:04.205 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:04.205 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:20:04.773 [2024-10-01 20:18:59.769017] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:20:04.773 "name": "raid_bdev1", 00:20:04.773 "aliases": [ 00:20:04.773 "2cfc7c6d-dd67-42c7-ad97-88ac1dcdf9e3" 00:20:04.773 ], 00:20:04.773 "product_name": "Raid Volume", 00:20:04.773 "block_size": 512, 00:20:04.773 "num_blocks": 126976, 00:20:04.773 "uuid": "2cfc7c6d-dd67-42c7-ad97-88ac1dcdf9e3", 00:20:04.773 "assigned_rate_limits": { 00:20:04.773 "rw_ios_per_sec": 0, 00:20:04.773 "rw_mbytes_per_sec": 0, 00:20:04.773 "r_mbytes_per_sec": 0, 00:20:04.773 "w_mbytes_per_sec": 0 00:20:04.773 }, 00:20:04.773 "claimed": false, 00:20:04.773 "zoned": false, 00:20:04.773 "supported_io_types": { 00:20:04.773 "read": true, 00:20:04.773 "write": true, 00:20:04.773 "unmap": true, 00:20:04.773 "flush": true, 00:20:04.773 "reset": true, 00:20:04.773 "nvme_admin": false, 00:20:04.773 "nvme_io": false, 00:20:04.773 "nvme_io_md": false, 00:20:04.773 "write_zeroes": true, 00:20:04.773 "zcopy": false, 00:20:04.773 "get_zone_info": false, 00:20:04.773 "zone_management": false, 00:20:04.773 "zone_append": false, 00:20:04.773 "compare": false, 00:20:04.773 "compare_and_write": false, 00:20:04.773 "abort": false, 00:20:04.773 "seek_hole": false, 00:20:04.773 "seek_data": false, 00:20:04.773 "copy": false, 00:20:04.773 "nvme_iov_md": false 00:20:04.773 }, 00:20:04.773 "memory_domains": [ 00:20:04.773 { 00:20:04.773 "dma_device_id": "system", 00:20:04.773 "dma_device_type": 1 00:20:04.773 }, 00:20:04.773 { 00:20:04.773 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:04.773 "dma_device_type": 2 00:20:04.773 }, 00:20:04.773 { 00:20:04.773 "dma_device_id": "system", 00:20:04.773 "dma_device_type": 1 00:20:04.773 }, 00:20:04.773 { 00:20:04.773 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:04.773 "dma_device_type": 2 00:20:04.773 } 00:20:04.773 ], 00:20:04.773 "driver_specific": { 00:20:04.773 "raid": { 00:20:04.773 "uuid": "2cfc7c6d-dd67-42c7-ad97-88ac1dcdf9e3", 00:20:04.773 "strip_size_kb": 64, 00:20:04.773 "state": "online", 00:20:04.773 "raid_level": "raid0", 00:20:04.773 "superblock": true, 00:20:04.773 "num_base_bdevs": 2, 00:20:04.773 "num_base_bdevs_discovered": 2, 00:20:04.773 "num_base_bdevs_operational": 2, 00:20:04.773 "base_bdevs_list": [ 00:20:04.773 { 00:20:04.773 "name": "pt1", 00:20:04.773 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:04.773 "is_configured": true, 00:20:04.773 "data_offset": 2048, 00:20:04.773 "data_size": 63488 00:20:04.773 }, 00:20:04.773 { 00:20:04.773 "name": "pt2", 00:20:04.773 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:04.773 "is_configured": true, 00:20:04.773 "data_offset": 2048, 00:20:04.773 "data_size": 63488 00:20:04.773 } 00:20:04.773 ] 00:20:04.773 } 00:20:04.773 } 00:20:04.773 }' 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:20:04.773 pt2' 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:04.773 20:18:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:04.773 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:04.773 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:04.773 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:20:04.773 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:20:04.773 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:04.773 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:04.773 [2024-10-01 20:19:00.017025] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:05.033 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:05.033 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=2cfc7c6d-dd67-42c7-ad97-88ac1dcdf9e3 00:20:05.033 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 2cfc7c6d-dd67-42c7-ad97-88ac1dcdf9e3 ']' 00:20:05.033 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:20:05.033 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:05.033 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:05.033 [2024-10-01 20:19:00.064686] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:20:05.034 [2024-10-01 20:19:00.064881] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:20:05.034 [2024-10-01 20:19:00.065109] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:05.034 [2024-10-01 20:19:00.065306] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:20:05.034 [2024-10-01 20:19:00.065451] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:05.034 [2024-10-01 20:19:00.196761] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:20:05.034 [2024-10-01 20:19:00.199589] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:20:05.034 [2024-10-01 20:19:00.199702] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:20:05.034 [2024-10-01 20:19:00.199794] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:20:05.034 [2024-10-01 20:19:00.199823] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:20:05.034 [2024-10-01 20:19:00.199841] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:20:05.034 request: 00:20:05.034 { 00:20:05.034 "name": "raid_bdev1", 00:20:05.034 "raid_level": "raid0", 00:20:05.034 "base_bdevs": [ 00:20:05.034 "malloc1", 00:20:05.034 "malloc2" 00:20:05.034 ], 00:20:05.034 "strip_size_kb": 64, 00:20:05.034 "superblock": false, 00:20:05.034 "method": "bdev_raid_create", 00:20:05.034 "req_id": 1 00:20:05.034 } 00:20:05.034 Got JSON-RPC error response 00:20:05.034 response: 00:20:05.034 { 00:20:05.034 "code": -17, 00:20:05.034 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:20:05.034 } 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:05.034 [2024-10-01 20:19:00.260751] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:20:05.034 [2024-10-01 20:19:00.260842] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:05.034 [2024-10-01 20:19:00.260882] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:20:05.034 [2024-10-01 20:19:00.260902] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:05.034 [2024-10-01 20:19:00.263972] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:05.034 [2024-10-01 20:19:00.264026] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:20:05.034 [2024-10-01 20:19:00.264142] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:20:05.034 [2024-10-01 20:19:00.264221] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:20:05.034 pt1 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 2 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:05.034 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:05.294 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:05.294 "name": "raid_bdev1", 00:20:05.294 "uuid": "2cfc7c6d-dd67-42c7-ad97-88ac1dcdf9e3", 00:20:05.294 "strip_size_kb": 64, 00:20:05.294 "state": "configuring", 00:20:05.294 "raid_level": "raid0", 00:20:05.294 "superblock": true, 00:20:05.294 "num_base_bdevs": 2, 00:20:05.294 "num_base_bdevs_discovered": 1, 00:20:05.294 "num_base_bdevs_operational": 2, 00:20:05.294 "base_bdevs_list": [ 00:20:05.294 { 00:20:05.294 "name": "pt1", 00:20:05.294 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:05.294 "is_configured": true, 00:20:05.294 "data_offset": 2048, 00:20:05.294 "data_size": 63488 00:20:05.294 }, 00:20:05.294 { 00:20:05.294 "name": null, 00:20:05.294 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:05.294 "is_configured": false, 00:20:05.294 "data_offset": 2048, 00:20:05.294 "data_size": 63488 00:20:05.294 } 00:20:05.294 ] 00:20:05.294 }' 00:20:05.294 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:05.294 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:05.553 [2024-10-01 20:19:00.780935] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:20:05.553 [2024-10-01 20:19:00.781042] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:05.553 [2024-10-01 20:19:00.781079] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:20:05.553 [2024-10-01 20:19:00.781099] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:05.553 [2024-10-01 20:19:00.781767] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:05.553 [2024-10-01 20:19:00.781820] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:20:05.553 [2024-10-01 20:19:00.781927] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:20:05.553 [2024-10-01 20:19:00.781964] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:20:05.553 [2024-10-01 20:19:00.782125] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:20:05.553 [2024-10-01 20:19:00.782161] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:20:05.553 [2024-10-01 20:19:00.782478] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:20:05.553 [2024-10-01 20:19:00.782675] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:20:05.553 [2024-10-01 20:19:00.782692] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:20:05.553 [2024-10-01 20:19:00.782907] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:05.553 pt2 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:05.553 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:05.812 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:05.812 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:05.812 "name": "raid_bdev1", 00:20:05.812 "uuid": "2cfc7c6d-dd67-42c7-ad97-88ac1dcdf9e3", 00:20:05.812 "strip_size_kb": 64, 00:20:05.812 "state": "online", 00:20:05.812 "raid_level": "raid0", 00:20:05.812 "superblock": true, 00:20:05.812 "num_base_bdevs": 2, 00:20:05.812 "num_base_bdevs_discovered": 2, 00:20:05.812 "num_base_bdevs_operational": 2, 00:20:05.812 "base_bdevs_list": [ 00:20:05.812 { 00:20:05.812 "name": "pt1", 00:20:05.812 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:05.812 "is_configured": true, 00:20:05.812 "data_offset": 2048, 00:20:05.812 "data_size": 63488 00:20:05.812 }, 00:20:05.812 { 00:20:05.812 "name": "pt2", 00:20:05.812 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:05.812 "is_configured": true, 00:20:05.812 "data_offset": 2048, 00:20:05.812 "data_size": 63488 00:20:05.812 } 00:20:05.812 ] 00:20:05.812 }' 00:20:05.812 20:19:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:05.812 20:19:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:06.071 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:20:06.071 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:20:06.071 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:20:06.071 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:20:06.071 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:20:06.071 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:20:06.071 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:20:06.071 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:06.071 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:06.071 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:20:06.071 [2024-10-01 20:19:01.317459] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:06.331 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:06.331 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:20:06.331 "name": "raid_bdev1", 00:20:06.331 "aliases": [ 00:20:06.331 "2cfc7c6d-dd67-42c7-ad97-88ac1dcdf9e3" 00:20:06.331 ], 00:20:06.331 "product_name": "Raid Volume", 00:20:06.332 "block_size": 512, 00:20:06.332 "num_blocks": 126976, 00:20:06.332 "uuid": "2cfc7c6d-dd67-42c7-ad97-88ac1dcdf9e3", 00:20:06.332 "assigned_rate_limits": { 00:20:06.332 "rw_ios_per_sec": 0, 00:20:06.332 "rw_mbytes_per_sec": 0, 00:20:06.332 "r_mbytes_per_sec": 0, 00:20:06.332 "w_mbytes_per_sec": 0 00:20:06.332 }, 00:20:06.332 "claimed": false, 00:20:06.332 "zoned": false, 00:20:06.332 "supported_io_types": { 00:20:06.332 "read": true, 00:20:06.332 "write": true, 00:20:06.332 "unmap": true, 00:20:06.332 "flush": true, 00:20:06.332 "reset": true, 00:20:06.332 "nvme_admin": false, 00:20:06.332 "nvme_io": false, 00:20:06.332 "nvme_io_md": false, 00:20:06.332 "write_zeroes": true, 00:20:06.332 "zcopy": false, 00:20:06.332 "get_zone_info": false, 00:20:06.332 "zone_management": false, 00:20:06.332 "zone_append": false, 00:20:06.332 "compare": false, 00:20:06.332 "compare_and_write": false, 00:20:06.332 "abort": false, 00:20:06.332 "seek_hole": false, 00:20:06.332 "seek_data": false, 00:20:06.332 "copy": false, 00:20:06.332 "nvme_iov_md": false 00:20:06.332 }, 00:20:06.332 "memory_domains": [ 00:20:06.332 { 00:20:06.332 "dma_device_id": "system", 00:20:06.332 "dma_device_type": 1 00:20:06.332 }, 00:20:06.332 { 00:20:06.332 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:06.332 "dma_device_type": 2 00:20:06.332 }, 00:20:06.332 { 00:20:06.332 "dma_device_id": "system", 00:20:06.332 "dma_device_type": 1 00:20:06.332 }, 00:20:06.332 { 00:20:06.332 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:06.332 "dma_device_type": 2 00:20:06.332 } 00:20:06.332 ], 00:20:06.332 "driver_specific": { 00:20:06.332 "raid": { 00:20:06.332 "uuid": "2cfc7c6d-dd67-42c7-ad97-88ac1dcdf9e3", 00:20:06.332 "strip_size_kb": 64, 00:20:06.332 "state": "online", 00:20:06.332 "raid_level": "raid0", 00:20:06.332 "superblock": true, 00:20:06.332 "num_base_bdevs": 2, 00:20:06.332 "num_base_bdevs_discovered": 2, 00:20:06.332 "num_base_bdevs_operational": 2, 00:20:06.332 "base_bdevs_list": [ 00:20:06.332 { 00:20:06.332 "name": "pt1", 00:20:06.332 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:06.332 "is_configured": true, 00:20:06.332 "data_offset": 2048, 00:20:06.332 "data_size": 63488 00:20:06.332 }, 00:20:06.332 { 00:20:06.332 "name": "pt2", 00:20:06.332 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:06.332 "is_configured": true, 00:20:06.332 "data_offset": 2048, 00:20:06.332 "data_size": 63488 00:20:06.332 } 00:20:06.332 ] 00:20:06.332 } 00:20:06.332 } 00:20:06.332 }' 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:20:06.332 pt2' 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:06.332 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:06.592 [2024-10-01 20:19:01.589535] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 2cfc7c6d-dd67-42c7-ad97-88ac1dcdf9e3 '!=' 2cfc7c6d-dd67-42c7-ad97-88ac1dcdf9e3 ']' 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 61542 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 61542 ']' 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 61542 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 61542 00:20:06.592 killing process with pid 61542 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 61542' 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 61542 00:20:06.592 [2024-10-01 20:19:01.667627] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:20:06.592 20:19:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 61542 00:20:06.592 [2024-10-01 20:19:01.667766] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:06.592 [2024-10-01 20:19:01.667838] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:20:06.592 [2024-10-01 20:19:01.667859] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:20:06.851 [2024-10-01 20:19:01.858583] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:20:08.756 20:19:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:20:08.756 00:20:08.756 real 0m5.652s 00:20:08.756 user 0m7.838s 00:20:08.756 sys 0m0.860s 00:20:08.756 ************************************ 00:20:08.756 END TEST raid_superblock_test 00:20:08.756 ************************************ 00:20:08.756 20:19:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:20:08.756 20:19:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:08.756 20:19:03 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 2 read 00:20:08.756 20:19:03 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:20:08.756 20:19:03 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:20:08.756 20:19:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:20:08.756 ************************************ 00:20:08.756 START TEST raid_read_error_test 00:20:08.756 ************************************ 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 2 read 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.8ND2GFQZXg 00:20:08.756 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=61770 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 61770 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 61770 ']' 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:08.756 20:19:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:08.756 [2024-10-01 20:19:03.775534] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:20:08.756 [2024-10-01 20:19:03.775707] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid61770 ] 00:20:08.756 [2024-10-01 20:19:03.945584] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:09.015 [2024-10-01 20:19:04.189888] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:20:09.274 [2024-10-01 20:19:04.396428] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:09.274 [2024-10-01 20:19:04.396477] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:09.841 BaseBdev1_malloc 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:09.841 true 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:09.841 [2024-10-01 20:19:04.862369] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:20:09.841 [2024-10-01 20:19:04.862455] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:09.841 [2024-10-01 20:19:04.862482] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:20:09.841 [2024-10-01 20:19:04.862501] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:09.841 [2024-10-01 20:19:04.865490] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:09.841 [2024-10-01 20:19:04.865572] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:20:09.841 BaseBdev1 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:09.841 BaseBdev2_malloc 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:09.841 true 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:09.841 [2024-10-01 20:19:04.920204] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:20:09.841 [2024-10-01 20:19:04.920288] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:09.841 [2024-10-01 20:19:04.920315] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:20:09.841 [2024-10-01 20:19:04.920334] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:09.841 [2024-10-01 20:19:04.923296] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:09.841 [2024-10-01 20:19:04.923361] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:20:09.841 BaseBdev2 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:09.841 [2024-10-01 20:19:04.928296] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:20:09.841 [2024-10-01 20:19:04.930883] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:20:09.841 [2024-10-01 20:19:04.931167] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:20:09.841 [2024-10-01 20:19:04.931190] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:20:09.841 [2024-10-01 20:19:04.931466] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:20:09.841 [2024-10-01 20:19:04.931660] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:20:09.841 [2024-10-01 20:19:04.931676] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:20:09.841 [2024-10-01 20:19:04.931917] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:09.841 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:09.841 "name": "raid_bdev1", 00:20:09.841 "uuid": "e960989f-289b-48a7-852e-737f9dcce663", 00:20:09.841 "strip_size_kb": 64, 00:20:09.841 "state": "online", 00:20:09.841 "raid_level": "raid0", 00:20:09.841 "superblock": true, 00:20:09.841 "num_base_bdevs": 2, 00:20:09.841 "num_base_bdevs_discovered": 2, 00:20:09.841 "num_base_bdevs_operational": 2, 00:20:09.841 "base_bdevs_list": [ 00:20:09.841 { 00:20:09.841 "name": "BaseBdev1", 00:20:09.841 "uuid": "b7fd09e5-8dcb-5e05-9430-9c1a11c5c82d", 00:20:09.841 "is_configured": true, 00:20:09.841 "data_offset": 2048, 00:20:09.841 "data_size": 63488 00:20:09.841 }, 00:20:09.841 { 00:20:09.841 "name": "BaseBdev2", 00:20:09.841 "uuid": "300f04da-056d-59d3-8b47-b5992a4220cb", 00:20:09.841 "is_configured": true, 00:20:09.841 "data_offset": 2048, 00:20:09.841 "data_size": 63488 00:20:09.841 } 00:20:09.842 ] 00:20:09.842 }' 00:20:09.842 20:19:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:09.842 20:19:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:10.409 20:19:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:20:10.409 20:19:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:20:10.409 [2024-10-01 20:19:05.574142] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:11.344 "name": "raid_bdev1", 00:20:11.344 "uuid": "e960989f-289b-48a7-852e-737f9dcce663", 00:20:11.344 "strip_size_kb": 64, 00:20:11.344 "state": "online", 00:20:11.344 "raid_level": "raid0", 00:20:11.344 "superblock": true, 00:20:11.344 "num_base_bdevs": 2, 00:20:11.344 "num_base_bdevs_discovered": 2, 00:20:11.344 "num_base_bdevs_operational": 2, 00:20:11.344 "base_bdevs_list": [ 00:20:11.344 { 00:20:11.344 "name": "BaseBdev1", 00:20:11.344 "uuid": "b7fd09e5-8dcb-5e05-9430-9c1a11c5c82d", 00:20:11.344 "is_configured": true, 00:20:11.344 "data_offset": 2048, 00:20:11.344 "data_size": 63488 00:20:11.344 }, 00:20:11.344 { 00:20:11.344 "name": "BaseBdev2", 00:20:11.344 "uuid": "300f04da-056d-59d3-8b47-b5992a4220cb", 00:20:11.344 "is_configured": true, 00:20:11.344 "data_offset": 2048, 00:20:11.344 "data_size": 63488 00:20:11.344 } 00:20:11.344 ] 00:20:11.344 }' 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:11.344 20:19:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:11.911 20:19:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:20:11.911 20:19:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:11.911 20:19:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:11.911 [2024-10-01 20:19:06.997377] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:20:11.911 [2024-10-01 20:19:06.997589] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:20:11.911 [2024-10-01 20:19:07.001295] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:11.911 [2024-10-01 20:19:07.001556] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:11.911 [2024-10-01 20:19:07.001650] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:20:11.911 [2024-10-01 20:19:07.001861] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:20:11.911 { 00:20:11.911 "results": [ 00:20:11.911 { 00:20:11.911 "job": "raid_bdev1", 00:20:11.911 "core_mask": "0x1", 00:20:11.911 "workload": "randrw", 00:20:11.911 "percentage": 50, 00:20:11.911 "status": "finished", 00:20:11.911 "queue_depth": 1, 00:20:11.911 "io_size": 131072, 00:20:11.911 "runtime": 1.420738, 00:20:11.911 "iops": 10635.317701082113, 00:20:11.911 "mibps": 1329.414712635264, 00:20:11.911 "io_failed": 1, 00:20:11.911 "io_timeout": 0, 00:20:11.911 "avg_latency_us": 131.40037323803853, 00:20:11.911 "min_latency_us": 39.33090909090909, 00:20:11.911 "max_latency_us": 1921.3963636363637 00:20:11.911 } 00:20:11.911 ], 00:20:11.911 "core_count": 1 00:20:11.911 } 00:20:11.911 20:19:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:11.911 20:19:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 61770 00:20:11.911 20:19:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 61770 ']' 00:20:11.911 20:19:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 61770 00:20:11.911 20:19:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:20:11.911 20:19:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:20:11.911 20:19:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 61770 00:20:11.911 killing process with pid 61770 00:20:11.911 20:19:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:20:11.911 20:19:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:20:11.911 20:19:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 61770' 00:20:11.911 20:19:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 61770 00:20:11.911 [2024-10-01 20:19:07.044317] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:20:11.911 20:19:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 61770 00:20:12.170 [2024-10-01 20:19:07.168480] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:20:14.081 20:19:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.8ND2GFQZXg 00:20:14.081 20:19:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:20:14.081 20:19:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:20:14.081 20:19:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.70 00:20:14.081 20:19:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:20:14.081 20:19:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:20:14.081 20:19:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:20:14.081 20:19:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.70 != \0\.\0\0 ]] 00:20:14.081 00:20:14.081 real 0m5.268s 00:20:14.081 user 0m6.339s 00:20:14.081 sys 0m0.676s 00:20:14.081 20:19:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:20:14.081 ************************************ 00:20:14.081 END TEST raid_read_error_test 00:20:14.081 ************************************ 00:20:14.081 20:19:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:14.081 20:19:08 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 2 write 00:20:14.081 20:19:08 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:20:14.081 20:19:08 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:20:14.081 20:19:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:20:14.081 ************************************ 00:20:14.081 START TEST raid_write_error_test 00:20:14.081 ************************************ 00:20:14.081 20:19:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 2 write 00:20:14.081 20:19:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:20:14.081 20:19:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:20:14.081 20:19:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:20:14.081 20:19:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:20:14.081 20:19:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:20:14.081 20:19:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:20:14.081 20:19:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:20:14.081 20:19:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:20:14.081 20:19:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:20:14.081 20:19:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:20:14.081 20:19:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:20:14.081 20:19:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:20:14.081 20:19:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:20:14.081 20:19:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:20:14.081 20:19:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:20:14.081 20:19:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:20:14.081 20:19:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:20:14.081 20:19:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:20:14.081 20:19:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:20:14.081 20:19:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:20:14.082 20:19:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:20:14.082 20:19:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:20:14.082 20:19:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.r6GCssZdgX 00:20:14.082 20:19:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=61921 00:20:14.082 20:19:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:20:14.082 20:19:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 61921 00:20:14.082 20:19:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 61921 ']' 00:20:14.082 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:14.082 20:19:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:14.082 20:19:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:14.082 20:19:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:14.082 20:19:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:14.082 20:19:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:14.082 [2024-10-01 20:19:09.119803] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:20:14.082 [2024-10-01 20:19:09.120338] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid61921 ] 00:20:14.082 [2024-10-01 20:19:09.302104] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:14.340 [2024-10-01 20:19:09.550336] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:20:14.598 [2024-10-01 20:19:09.756498] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:14.598 [2024-10-01 20:19:09.756586] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:15.165 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:15.165 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:20:15.165 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:20:15.165 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:15.166 BaseBdev1_malloc 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:15.166 true 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:15.166 [2024-10-01 20:19:10.218886] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:20:15.166 [2024-10-01 20:19:10.218972] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:15.166 [2024-10-01 20:19:10.219000] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:20:15.166 [2024-10-01 20:19:10.219019] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:15.166 [2024-10-01 20:19:10.222188] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:15.166 [2024-10-01 20:19:10.222252] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:20:15.166 BaseBdev1 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:15.166 BaseBdev2_malloc 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:15.166 true 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:15.166 [2024-10-01 20:19:10.274049] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:20:15.166 [2024-10-01 20:19:10.274133] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:15.166 [2024-10-01 20:19:10.274160] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:20:15.166 [2024-10-01 20:19:10.274179] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:15.166 [2024-10-01 20:19:10.277137] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:15.166 [2024-10-01 20:19:10.277370] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:20:15.166 BaseBdev2 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:15.166 [2024-10-01 20:19:10.282295] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:20:15.166 [2024-10-01 20:19:10.285008] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:20:15.166 [2024-10-01 20:19:10.285302] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:20:15.166 [2024-10-01 20:19:10.285344] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:20:15.166 [2024-10-01 20:19:10.285641] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:20:15.166 [2024-10-01 20:19:10.285919] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:20:15.166 [2024-10-01 20:19:10.285936] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:20:15.166 [2024-10-01 20:19:10.286235] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:15.166 "name": "raid_bdev1", 00:20:15.166 "uuid": "23bbf641-9203-434e-a250-3ea013bc3995", 00:20:15.166 "strip_size_kb": 64, 00:20:15.166 "state": "online", 00:20:15.166 "raid_level": "raid0", 00:20:15.166 "superblock": true, 00:20:15.166 "num_base_bdevs": 2, 00:20:15.166 "num_base_bdevs_discovered": 2, 00:20:15.166 "num_base_bdevs_operational": 2, 00:20:15.166 "base_bdevs_list": [ 00:20:15.166 { 00:20:15.166 "name": "BaseBdev1", 00:20:15.166 "uuid": "a71b42fa-2533-5456-a733-72bb1e43ea3a", 00:20:15.166 "is_configured": true, 00:20:15.166 "data_offset": 2048, 00:20:15.166 "data_size": 63488 00:20:15.166 }, 00:20:15.166 { 00:20:15.166 "name": "BaseBdev2", 00:20:15.166 "uuid": "304ccccf-5062-5f18-8c8a-127a79474180", 00:20:15.166 "is_configured": true, 00:20:15.166 "data_offset": 2048, 00:20:15.166 "data_size": 63488 00:20:15.166 } 00:20:15.166 ] 00:20:15.166 }' 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:15.166 20:19:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:15.733 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:20:15.733 20:19:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:20:15.733 [2024-10-01 20:19:10.927892] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:16.674 "name": "raid_bdev1", 00:20:16.674 "uuid": "23bbf641-9203-434e-a250-3ea013bc3995", 00:20:16.674 "strip_size_kb": 64, 00:20:16.674 "state": "online", 00:20:16.674 "raid_level": "raid0", 00:20:16.674 "superblock": true, 00:20:16.674 "num_base_bdevs": 2, 00:20:16.674 "num_base_bdevs_discovered": 2, 00:20:16.674 "num_base_bdevs_operational": 2, 00:20:16.674 "base_bdevs_list": [ 00:20:16.674 { 00:20:16.674 "name": "BaseBdev1", 00:20:16.674 "uuid": "a71b42fa-2533-5456-a733-72bb1e43ea3a", 00:20:16.674 "is_configured": true, 00:20:16.674 "data_offset": 2048, 00:20:16.674 "data_size": 63488 00:20:16.674 }, 00:20:16.674 { 00:20:16.674 "name": "BaseBdev2", 00:20:16.674 "uuid": "304ccccf-5062-5f18-8c8a-127a79474180", 00:20:16.674 "is_configured": true, 00:20:16.674 "data_offset": 2048, 00:20:16.674 "data_size": 63488 00:20:16.674 } 00:20:16.674 ] 00:20:16.674 }' 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:16.674 20:19:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:17.242 20:19:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:20:17.242 20:19:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:17.242 20:19:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:17.242 [2024-10-01 20:19:12.358022] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:20:17.242 [2024-10-01 20:19:12.358304] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:20:17.242 [2024-10-01 20:19:12.361863] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:17.242 [2024-10-01 20:19:12.361921] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:17.242 [2024-10-01 20:19:12.361966] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:20:17.242 [2024-10-01 20:19:12.361995] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:20:17.242 { 00:20:17.242 "results": [ 00:20:17.242 { 00:20:17.242 "job": "raid_bdev1", 00:20:17.242 "core_mask": "0x1", 00:20:17.242 "workload": "randrw", 00:20:17.242 "percentage": 50, 00:20:17.242 "status": "finished", 00:20:17.242 "queue_depth": 1, 00:20:17.242 "io_size": 131072, 00:20:17.242 "runtime": 1.427903, 00:20:17.242 "iops": 10467.09755494596, 00:20:17.242 "mibps": 1308.387194368245, 00:20:17.242 "io_failed": 1, 00:20:17.242 "io_timeout": 0, 00:20:17.242 "avg_latency_us": 133.23784280214332, 00:20:17.242 "min_latency_us": 39.33090909090909, 00:20:17.242 "max_latency_us": 1861.8181818181818 00:20:17.242 } 00:20:17.242 ], 00:20:17.242 "core_count": 1 00:20:17.242 } 00:20:17.242 20:19:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:17.242 20:19:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 61921 00:20:17.242 20:19:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 61921 ']' 00:20:17.242 20:19:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 61921 00:20:17.242 20:19:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:20:17.242 20:19:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:20:17.242 20:19:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 61921 00:20:17.242 killing process with pid 61921 00:20:17.242 20:19:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:20:17.242 20:19:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:20:17.242 20:19:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 61921' 00:20:17.242 20:19:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 61921 00:20:17.242 [2024-10-01 20:19:12.402364] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:20:17.242 20:19:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 61921 00:20:17.501 [2024-10-01 20:19:12.520749] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:20:19.403 20:19:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.r6GCssZdgX 00:20:19.403 20:19:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:20:19.403 20:19:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:20:19.403 20:19:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.70 00:20:19.403 20:19:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:20:19.403 ************************************ 00:20:19.403 END TEST raid_write_error_test 00:20:19.403 ************************************ 00:20:19.403 20:19:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:20:19.403 20:19:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:20:19.403 20:19:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.70 != \0\.\0\0 ]] 00:20:19.403 00:20:19.403 real 0m5.343s 00:20:19.403 user 0m6.370s 00:20:19.403 sys 0m0.690s 00:20:19.403 20:19:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:20:19.403 20:19:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:19.403 20:19:14 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:20:19.403 20:19:14 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 2 false 00:20:19.403 20:19:14 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:20:19.403 20:19:14 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:20:19.403 20:19:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:20:19.403 ************************************ 00:20:19.403 START TEST raid_state_function_test 00:20:19.403 ************************************ 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 2 false 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:20:19.403 Process raid pid: 62070 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=62070 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 62070' 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 62070 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 62070 ']' 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:19.403 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:19.403 20:19:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:19.403 [2024-10-01 20:19:14.520926] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:20:19.403 [2024-10-01 20:19:14.521158] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:19.660 [2024-10-01 20:19:14.700908] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:19.918 [2024-10-01 20:19:14.937290] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:20:19.918 [2024-10-01 20:19:15.136919] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:19.918 [2024-10-01 20:19:15.137025] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:20.526 20:19:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:20.526 20:19:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:20:20.526 20:19:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:20:20.526 20:19:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:20.526 20:19:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:20.526 [2024-10-01 20:19:15.553912] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:20:20.526 [2024-10-01 20:19:15.554162] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:20:20.526 [2024-10-01 20:19:15.554190] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:20:20.526 [2024-10-01 20:19:15.554210] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:20:20.526 20:19:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:20.526 20:19:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:20:20.526 20:19:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:20.526 20:19:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:20:20.526 20:19:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:20.526 20:19:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:20.526 20:19:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:20.526 20:19:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:20.526 20:19:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:20.526 20:19:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:20.527 20:19:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:20.527 20:19:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:20.527 20:19:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:20.527 20:19:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:20.527 20:19:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:20.527 20:19:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:20.527 20:19:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:20.527 "name": "Existed_Raid", 00:20:20.527 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:20.527 "strip_size_kb": 64, 00:20:20.527 "state": "configuring", 00:20:20.527 "raid_level": "concat", 00:20:20.527 "superblock": false, 00:20:20.527 "num_base_bdevs": 2, 00:20:20.527 "num_base_bdevs_discovered": 0, 00:20:20.527 "num_base_bdevs_operational": 2, 00:20:20.527 "base_bdevs_list": [ 00:20:20.527 { 00:20:20.527 "name": "BaseBdev1", 00:20:20.527 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:20.527 "is_configured": false, 00:20:20.527 "data_offset": 0, 00:20:20.527 "data_size": 0 00:20:20.527 }, 00:20:20.527 { 00:20:20.527 "name": "BaseBdev2", 00:20:20.527 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:20.527 "is_configured": false, 00:20:20.527 "data_offset": 0, 00:20:20.527 "data_size": 0 00:20:20.527 } 00:20:20.527 ] 00:20:20.527 }' 00:20:20.527 20:19:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:20.527 20:19:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:21.106 [2024-10-01 20:19:16.090008] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:20:21.106 [2024-10-01 20:19:16.090195] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:21.106 [2024-10-01 20:19:16.098034] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:20:21.106 [2024-10-01 20:19:16.098134] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:20:21.106 [2024-10-01 20:19:16.098149] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:20:21.106 [2024-10-01 20:19:16.098184] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:21.106 [2024-10-01 20:19:16.143793] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:20:21.106 BaseBdev1 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:21.106 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:21.106 [ 00:20:21.106 { 00:20:21.106 "name": "BaseBdev1", 00:20:21.106 "aliases": [ 00:20:21.106 "13b6ae5a-a43c-4efc-a6f7-87e638a5ded9" 00:20:21.106 ], 00:20:21.106 "product_name": "Malloc disk", 00:20:21.106 "block_size": 512, 00:20:21.106 "num_blocks": 65536, 00:20:21.106 "uuid": "13b6ae5a-a43c-4efc-a6f7-87e638a5ded9", 00:20:21.106 "assigned_rate_limits": { 00:20:21.106 "rw_ios_per_sec": 0, 00:20:21.106 "rw_mbytes_per_sec": 0, 00:20:21.106 "r_mbytes_per_sec": 0, 00:20:21.106 "w_mbytes_per_sec": 0 00:20:21.106 }, 00:20:21.106 "claimed": true, 00:20:21.106 "claim_type": "exclusive_write", 00:20:21.106 "zoned": false, 00:20:21.106 "supported_io_types": { 00:20:21.106 "read": true, 00:20:21.106 "write": true, 00:20:21.106 "unmap": true, 00:20:21.106 "flush": true, 00:20:21.106 "reset": true, 00:20:21.107 "nvme_admin": false, 00:20:21.107 "nvme_io": false, 00:20:21.107 "nvme_io_md": false, 00:20:21.107 "write_zeroes": true, 00:20:21.107 "zcopy": true, 00:20:21.107 "get_zone_info": false, 00:20:21.107 "zone_management": false, 00:20:21.107 "zone_append": false, 00:20:21.107 "compare": false, 00:20:21.107 "compare_and_write": false, 00:20:21.107 "abort": true, 00:20:21.107 "seek_hole": false, 00:20:21.107 "seek_data": false, 00:20:21.107 "copy": true, 00:20:21.107 "nvme_iov_md": false 00:20:21.107 }, 00:20:21.107 "memory_domains": [ 00:20:21.107 { 00:20:21.107 "dma_device_id": "system", 00:20:21.107 "dma_device_type": 1 00:20:21.107 }, 00:20:21.107 { 00:20:21.107 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:21.107 "dma_device_type": 2 00:20:21.107 } 00:20:21.107 ], 00:20:21.107 "driver_specific": {} 00:20:21.107 } 00:20:21.107 ] 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:21.107 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:21.107 "name": "Existed_Raid", 00:20:21.108 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:21.108 "strip_size_kb": 64, 00:20:21.108 "state": "configuring", 00:20:21.108 "raid_level": "concat", 00:20:21.108 "superblock": false, 00:20:21.108 "num_base_bdevs": 2, 00:20:21.108 "num_base_bdevs_discovered": 1, 00:20:21.108 "num_base_bdevs_operational": 2, 00:20:21.108 "base_bdevs_list": [ 00:20:21.108 { 00:20:21.108 "name": "BaseBdev1", 00:20:21.108 "uuid": "13b6ae5a-a43c-4efc-a6f7-87e638a5ded9", 00:20:21.108 "is_configured": true, 00:20:21.108 "data_offset": 0, 00:20:21.108 "data_size": 65536 00:20:21.108 }, 00:20:21.108 { 00:20:21.108 "name": "BaseBdev2", 00:20:21.108 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:21.108 "is_configured": false, 00:20:21.108 "data_offset": 0, 00:20:21.108 "data_size": 0 00:20:21.108 } 00:20:21.108 ] 00:20:21.108 }' 00:20:21.108 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:21.108 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:21.689 [2024-10-01 20:19:16.736068] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:20:21.689 [2024-10-01 20:19:16.736185] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:21.689 [2024-10-01 20:19:16.748079] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:20:21.689 [2024-10-01 20:19:16.750916] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:20:21.689 [2024-10-01 20:19:16.750974] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:21.689 "name": "Existed_Raid", 00:20:21.689 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:21.689 "strip_size_kb": 64, 00:20:21.689 "state": "configuring", 00:20:21.689 "raid_level": "concat", 00:20:21.689 "superblock": false, 00:20:21.689 "num_base_bdevs": 2, 00:20:21.689 "num_base_bdevs_discovered": 1, 00:20:21.689 "num_base_bdevs_operational": 2, 00:20:21.689 "base_bdevs_list": [ 00:20:21.689 { 00:20:21.689 "name": "BaseBdev1", 00:20:21.689 "uuid": "13b6ae5a-a43c-4efc-a6f7-87e638a5ded9", 00:20:21.689 "is_configured": true, 00:20:21.689 "data_offset": 0, 00:20:21.689 "data_size": 65536 00:20:21.689 }, 00:20:21.689 { 00:20:21.689 "name": "BaseBdev2", 00:20:21.689 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:21.689 "is_configured": false, 00:20:21.689 "data_offset": 0, 00:20:21.689 "data_size": 0 00:20:21.689 } 00:20:21.689 ] 00:20:21.689 }' 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:21.689 20:19:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:22.256 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:20:22.256 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:22.256 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:22.256 [2024-10-01 20:19:17.307589] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:20:22.256 [2024-10-01 20:19:17.307656] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:20:22.256 [2024-10-01 20:19:17.307670] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:20:22.256 [2024-10-01 20:19:17.308102] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:20:22.256 [2024-10-01 20:19:17.308320] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:20:22.256 [2024-10-01 20:19:17.308341] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:20:22.256 [2024-10-01 20:19:17.308661] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:22.256 BaseBdev2 00:20:22.256 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:22.256 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:20:22.256 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:20:22.256 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:20:22.256 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:22.257 [ 00:20:22.257 { 00:20:22.257 "name": "BaseBdev2", 00:20:22.257 "aliases": [ 00:20:22.257 "aa5ead8d-e804-44e0-9a26-438750e6aad6" 00:20:22.257 ], 00:20:22.257 "product_name": "Malloc disk", 00:20:22.257 "block_size": 512, 00:20:22.257 "num_blocks": 65536, 00:20:22.257 "uuid": "aa5ead8d-e804-44e0-9a26-438750e6aad6", 00:20:22.257 "assigned_rate_limits": { 00:20:22.257 "rw_ios_per_sec": 0, 00:20:22.257 "rw_mbytes_per_sec": 0, 00:20:22.257 "r_mbytes_per_sec": 0, 00:20:22.257 "w_mbytes_per_sec": 0 00:20:22.257 }, 00:20:22.257 "claimed": true, 00:20:22.257 "claim_type": "exclusive_write", 00:20:22.257 "zoned": false, 00:20:22.257 "supported_io_types": { 00:20:22.257 "read": true, 00:20:22.257 "write": true, 00:20:22.257 "unmap": true, 00:20:22.257 "flush": true, 00:20:22.257 "reset": true, 00:20:22.257 "nvme_admin": false, 00:20:22.257 "nvme_io": false, 00:20:22.257 "nvme_io_md": false, 00:20:22.257 "write_zeroes": true, 00:20:22.257 "zcopy": true, 00:20:22.257 "get_zone_info": false, 00:20:22.257 "zone_management": false, 00:20:22.257 "zone_append": false, 00:20:22.257 "compare": false, 00:20:22.257 "compare_and_write": false, 00:20:22.257 "abort": true, 00:20:22.257 "seek_hole": false, 00:20:22.257 "seek_data": false, 00:20:22.257 "copy": true, 00:20:22.257 "nvme_iov_md": false 00:20:22.257 }, 00:20:22.257 "memory_domains": [ 00:20:22.257 { 00:20:22.257 "dma_device_id": "system", 00:20:22.257 "dma_device_type": 1 00:20:22.257 }, 00:20:22.257 { 00:20:22.257 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:22.257 "dma_device_type": 2 00:20:22.257 } 00:20:22.257 ], 00:20:22.257 "driver_specific": {} 00:20:22.257 } 00:20:22.257 ] 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:22.257 "name": "Existed_Raid", 00:20:22.257 "uuid": "0abfcca8-3dbd-4e5e-b8a0-179c70f689dd", 00:20:22.257 "strip_size_kb": 64, 00:20:22.257 "state": "online", 00:20:22.257 "raid_level": "concat", 00:20:22.257 "superblock": false, 00:20:22.257 "num_base_bdevs": 2, 00:20:22.257 "num_base_bdevs_discovered": 2, 00:20:22.257 "num_base_bdevs_operational": 2, 00:20:22.257 "base_bdevs_list": [ 00:20:22.257 { 00:20:22.257 "name": "BaseBdev1", 00:20:22.257 "uuid": "13b6ae5a-a43c-4efc-a6f7-87e638a5ded9", 00:20:22.257 "is_configured": true, 00:20:22.257 "data_offset": 0, 00:20:22.257 "data_size": 65536 00:20:22.257 }, 00:20:22.257 { 00:20:22.257 "name": "BaseBdev2", 00:20:22.257 "uuid": "aa5ead8d-e804-44e0-9a26-438750e6aad6", 00:20:22.257 "is_configured": true, 00:20:22.257 "data_offset": 0, 00:20:22.257 "data_size": 65536 00:20:22.257 } 00:20:22.257 ] 00:20:22.257 }' 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:22.257 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:22.823 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:20:22.823 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:20:22.823 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:20:22.823 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:20:22.823 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:20:22.823 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:20:22.823 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:20:22.823 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:20:22.823 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:22.823 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:22.823 [2024-10-01 20:19:17.864244] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:22.823 20:19:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:22.823 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:20:22.823 "name": "Existed_Raid", 00:20:22.823 "aliases": [ 00:20:22.823 "0abfcca8-3dbd-4e5e-b8a0-179c70f689dd" 00:20:22.823 ], 00:20:22.823 "product_name": "Raid Volume", 00:20:22.823 "block_size": 512, 00:20:22.823 "num_blocks": 131072, 00:20:22.823 "uuid": "0abfcca8-3dbd-4e5e-b8a0-179c70f689dd", 00:20:22.823 "assigned_rate_limits": { 00:20:22.823 "rw_ios_per_sec": 0, 00:20:22.823 "rw_mbytes_per_sec": 0, 00:20:22.823 "r_mbytes_per_sec": 0, 00:20:22.823 "w_mbytes_per_sec": 0 00:20:22.823 }, 00:20:22.823 "claimed": false, 00:20:22.823 "zoned": false, 00:20:22.823 "supported_io_types": { 00:20:22.823 "read": true, 00:20:22.823 "write": true, 00:20:22.823 "unmap": true, 00:20:22.823 "flush": true, 00:20:22.823 "reset": true, 00:20:22.823 "nvme_admin": false, 00:20:22.823 "nvme_io": false, 00:20:22.823 "nvme_io_md": false, 00:20:22.823 "write_zeroes": true, 00:20:22.823 "zcopy": false, 00:20:22.823 "get_zone_info": false, 00:20:22.823 "zone_management": false, 00:20:22.823 "zone_append": false, 00:20:22.823 "compare": false, 00:20:22.823 "compare_and_write": false, 00:20:22.823 "abort": false, 00:20:22.823 "seek_hole": false, 00:20:22.823 "seek_data": false, 00:20:22.823 "copy": false, 00:20:22.823 "nvme_iov_md": false 00:20:22.823 }, 00:20:22.823 "memory_domains": [ 00:20:22.823 { 00:20:22.823 "dma_device_id": "system", 00:20:22.823 "dma_device_type": 1 00:20:22.823 }, 00:20:22.823 { 00:20:22.823 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:22.823 "dma_device_type": 2 00:20:22.823 }, 00:20:22.823 { 00:20:22.823 "dma_device_id": "system", 00:20:22.823 "dma_device_type": 1 00:20:22.823 }, 00:20:22.823 { 00:20:22.823 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:22.823 "dma_device_type": 2 00:20:22.823 } 00:20:22.823 ], 00:20:22.823 "driver_specific": { 00:20:22.823 "raid": { 00:20:22.823 "uuid": "0abfcca8-3dbd-4e5e-b8a0-179c70f689dd", 00:20:22.823 "strip_size_kb": 64, 00:20:22.823 "state": "online", 00:20:22.823 "raid_level": "concat", 00:20:22.823 "superblock": false, 00:20:22.823 "num_base_bdevs": 2, 00:20:22.823 "num_base_bdevs_discovered": 2, 00:20:22.823 "num_base_bdevs_operational": 2, 00:20:22.823 "base_bdevs_list": [ 00:20:22.823 { 00:20:22.823 "name": "BaseBdev1", 00:20:22.823 "uuid": "13b6ae5a-a43c-4efc-a6f7-87e638a5ded9", 00:20:22.823 "is_configured": true, 00:20:22.823 "data_offset": 0, 00:20:22.823 "data_size": 65536 00:20:22.823 }, 00:20:22.823 { 00:20:22.823 "name": "BaseBdev2", 00:20:22.823 "uuid": "aa5ead8d-e804-44e0-9a26-438750e6aad6", 00:20:22.823 "is_configured": true, 00:20:22.823 "data_offset": 0, 00:20:22.823 "data_size": 65536 00:20:22.823 } 00:20:22.823 ] 00:20:22.823 } 00:20:22.823 } 00:20:22.823 }' 00:20:22.823 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:20:22.823 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:20:22.823 BaseBdev2' 00:20:22.823 20:19:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:22.823 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:20:22.823 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:22.823 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:20:22.823 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:22.823 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:22.824 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:22.824 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:23.082 [2024-10-01 20:19:18.139994] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:20:23.082 [2024-10-01 20:19:18.140039] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:20:23.082 [2024-10-01 20:19:18.140158] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:23.082 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:23.082 "name": "Existed_Raid", 00:20:23.082 "uuid": "0abfcca8-3dbd-4e5e-b8a0-179c70f689dd", 00:20:23.082 "strip_size_kb": 64, 00:20:23.082 "state": "offline", 00:20:23.082 "raid_level": "concat", 00:20:23.082 "superblock": false, 00:20:23.082 "num_base_bdevs": 2, 00:20:23.082 "num_base_bdevs_discovered": 1, 00:20:23.082 "num_base_bdevs_operational": 1, 00:20:23.082 "base_bdevs_list": [ 00:20:23.082 { 00:20:23.083 "name": null, 00:20:23.083 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:23.083 "is_configured": false, 00:20:23.083 "data_offset": 0, 00:20:23.083 "data_size": 65536 00:20:23.083 }, 00:20:23.083 { 00:20:23.083 "name": "BaseBdev2", 00:20:23.083 "uuid": "aa5ead8d-e804-44e0-9a26-438750e6aad6", 00:20:23.083 "is_configured": true, 00:20:23.083 "data_offset": 0, 00:20:23.083 "data_size": 65536 00:20:23.083 } 00:20:23.083 ] 00:20:23.083 }' 00:20:23.083 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:23.083 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:23.649 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:20:23.649 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:20:23.649 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:23.649 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:20:23.649 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:23.649 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:23.649 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:23.649 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:20:23.649 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:20:23.649 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:20:23.649 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:23.649 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:23.649 [2024-10-01 20:19:18.816543] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:20:23.649 [2024-10-01 20:19:18.816829] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 62070 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 62070 ']' 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 62070 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 62070 00:20:23.908 killing process with pid 62070 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 62070' 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 62070 00:20:23.908 [2024-10-01 20:19:18.995834] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:20:23.908 20:19:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 62070 00:20:23.908 [2024-10-01 20:19:19.010744] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:20:25.808 20:19:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:20:25.808 00:20:25.808 real 0m6.366s 00:20:25.808 user 0m9.061s 00:20:25.808 sys 0m0.959s 00:20:25.808 20:19:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:20:25.808 20:19:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:25.808 ************************************ 00:20:25.808 END TEST raid_state_function_test 00:20:25.808 ************************************ 00:20:25.808 20:19:20 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 2 true 00:20:25.808 20:19:20 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:20:25.808 20:19:20 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:20:25.808 20:19:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:20:25.808 ************************************ 00:20:25.808 START TEST raid_state_function_test_sb 00:20:25.808 ************************************ 00:20:25.808 20:19:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 2 true 00:20:25.808 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:20:25.808 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:20:25.808 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:20:25.808 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:20:25.808 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:20:25.808 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:20:25.809 Process raid pid: 62340 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=62340 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 62340' 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 62340 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 62340 ']' 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:25.809 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:25.809 20:19:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:25.809 [2024-10-01 20:19:20.912184] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:20:25.809 [2024-10-01 20:19:20.912661] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:26.068 [2024-10-01 20:19:21.077804] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:26.326 [2024-10-01 20:19:21.324171] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:20:26.326 [2024-10-01 20:19:21.520338] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:26.326 [2024-10-01 20:19:21.520398] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:26.892 [2024-10-01 20:19:21.942641] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:20:26.892 [2024-10-01 20:19:21.942893] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:20:26.892 [2024-10-01 20:19:21.942929] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:20:26.892 [2024-10-01 20:19:21.942952] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:26.892 "name": "Existed_Raid", 00:20:26.892 "uuid": "620a961c-7cb8-4e58-8cbe-af5590323f9d", 00:20:26.892 "strip_size_kb": 64, 00:20:26.892 "state": "configuring", 00:20:26.892 "raid_level": "concat", 00:20:26.892 "superblock": true, 00:20:26.892 "num_base_bdevs": 2, 00:20:26.892 "num_base_bdevs_discovered": 0, 00:20:26.892 "num_base_bdevs_operational": 2, 00:20:26.892 "base_bdevs_list": [ 00:20:26.892 { 00:20:26.892 "name": "BaseBdev1", 00:20:26.892 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:26.892 "is_configured": false, 00:20:26.892 "data_offset": 0, 00:20:26.892 "data_size": 0 00:20:26.892 }, 00:20:26.892 { 00:20:26.892 "name": "BaseBdev2", 00:20:26.892 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:26.892 "is_configured": false, 00:20:26.892 "data_offset": 0, 00:20:26.892 "data_size": 0 00:20:26.892 } 00:20:26.892 ] 00:20:26.892 }' 00:20:26.892 20:19:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:26.892 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:27.459 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:20:27.459 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:27.459 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:27.459 [2024-10-01 20:19:22.482682] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:20:27.459 [2024-10-01 20:19:22.482766] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:20:27.459 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:27.459 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:20:27.459 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:27.460 [2024-10-01 20:19:22.490684] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:20:27.460 [2024-10-01 20:19:22.490908] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:20:27.460 [2024-10-01 20:19:22.490936] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:20:27.460 [2024-10-01 20:19:22.490960] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:27.460 [2024-10-01 20:19:22.536333] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:20:27.460 BaseBdev1 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:27.460 [ 00:20:27.460 { 00:20:27.460 "name": "BaseBdev1", 00:20:27.460 "aliases": [ 00:20:27.460 "165a0c04-851a-4ae4-aafb-2ae1fe2a8749" 00:20:27.460 ], 00:20:27.460 "product_name": "Malloc disk", 00:20:27.460 "block_size": 512, 00:20:27.460 "num_blocks": 65536, 00:20:27.460 "uuid": "165a0c04-851a-4ae4-aafb-2ae1fe2a8749", 00:20:27.460 "assigned_rate_limits": { 00:20:27.460 "rw_ios_per_sec": 0, 00:20:27.460 "rw_mbytes_per_sec": 0, 00:20:27.460 "r_mbytes_per_sec": 0, 00:20:27.460 "w_mbytes_per_sec": 0 00:20:27.460 }, 00:20:27.460 "claimed": true, 00:20:27.460 "claim_type": "exclusive_write", 00:20:27.460 "zoned": false, 00:20:27.460 "supported_io_types": { 00:20:27.460 "read": true, 00:20:27.460 "write": true, 00:20:27.460 "unmap": true, 00:20:27.460 "flush": true, 00:20:27.460 "reset": true, 00:20:27.460 "nvme_admin": false, 00:20:27.460 "nvme_io": false, 00:20:27.460 "nvme_io_md": false, 00:20:27.460 "write_zeroes": true, 00:20:27.460 "zcopy": true, 00:20:27.460 "get_zone_info": false, 00:20:27.460 "zone_management": false, 00:20:27.460 "zone_append": false, 00:20:27.460 "compare": false, 00:20:27.460 "compare_and_write": false, 00:20:27.460 "abort": true, 00:20:27.460 "seek_hole": false, 00:20:27.460 "seek_data": false, 00:20:27.460 "copy": true, 00:20:27.460 "nvme_iov_md": false 00:20:27.460 }, 00:20:27.460 "memory_domains": [ 00:20:27.460 { 00:20:27.460 "dma_device_id": "system", 00:20:27.460 "dma_device_type": 1 00:20:27.460 }, 00:20:27.460 { 00:20:27.460 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:27.460 "dma_device_type": 2 00:20:27.460 } 00:20:27.460 ], 00:20:27.460 "driver_specific": {} 00:20:27.460 } 00:20:27.460 ] 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:27.460 "name": "Existed_Raid", 00:20:27.460 "uuid": "f6d59e08-f728-4eb3-a243-679f03238a6c", 00:20:27.460 "strip_size_kb": 64, 00:20:27.460 "state": "configuring", 00:20:27.460 "raid_level": "concat", 00:20:27.460 "superblock": true, 00:20:27.460 "num_base_bdevs": 2, 00:20:27.460 "num_base_bdevs_discovered": 1, 00:20:27.460 "num_base_bdevs_operational": 2, 00:20:27.460 "base_bdevs_list": [ 00:20:27.460 { 00:20:27.460 "name": "BaseBdev1", 00:20:27.460 "uuid": "165a0c04-851a-4ae4-aafb-2ae1fe2a8749", 00:20:27.460 "is_configured": true, 00:20:27.460 "data_offset": 2048, 00:20:27.460 "data_size": 63488 00:20:27.460 }, 00:20:27.460 { 00:20:27.460 "name": "BaseBdev2", 00:20:27.460 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:27.460 "is_configured": false, 00:20:27.460 "data_offset": 0, 00:20:27.460 "data_size": 0 00:20:27.460 } 00:20:27.460 ] 00:20:27.460 }' 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:27.460 20:19:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:28.040 [2024-10-01 20:19:23.092532] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:20:28.040 [2024-10-01 20:19:23.092602] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:28.040 [2024-10-01 20:19:23.100555] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:20:28.040 [2024-10-01 20:19:23.103298] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:20:28.040 [2024-10-01 20:19:23.103485] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:28.040 "name": "Existed_Raid", 00:20:28.040 "uuid": "828f486c-5b7b-415a-9c26-a4f16a0fe496", 00:20:28.040 "strip_size_kb": 64, 00:20:28.040 "state": "configuring", 00:20:28.040 "raid_level": "concat", 00:20:28.040 "superblock": true, 00:20:28.040 "num_base_bdevs": 2, 00:20:28.040 "num_base_bdevs_discovered": 1, 00:20:28.040 "num_base_bdevs_operational": 2, 00:20:28.040 "base_bdevs_list": [ 00:20:28.040 { 00:20:28.040 "name": "BaseBdev1", 00:20:28.040 "uuid": "165a0c04-851a-4ae4-aafb-2ae1fe2a8749", 00:20:28.040 "is_configured": true, 00:20:28.040 "data_offset": 2048, 00:20:28.040 "data_size": 63488 00:20:28.040 }, 00:20:28.040 { 00:20:28.040 "name": "BaseBdev2", 00:20:28.040 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:28.040 "is_configured": false, 00:20:28.040 "data_offset": 0, 00:20:28.040 "data_size": 0 00:20:28.040 } 00:20:28.040 ] 00:20:28.040 }' 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:28.040 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:28.613 [2024-10-01 20:19:23.627590] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:20:28.613 [2024-10-01 20:19:23.628210] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:20:28.613 [2024-10-01 20:19:23.628241] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:20:28.613 BaseBdev2 00:20:28.613 [2024-10-01 20:19:23.628592] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:20:28.613 [2024-10-01 20:19:23.628812] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:20:28.613 [2024-10-01 20:19:23.628836] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:20:28.613 [2024-10-01 20:19:23.629024] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:28.613 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:28.613 [ 00:20:28.613 { 00:20:28.613 "name": "BaseBdev2", 00:20:28.613 "aliases": [ 00:20:28.613 "aa941b85-da99-4820-a3d2-13b6a77a4bb1" 00:20:28.613 ], 00:20:28.613 "product_name": "Malloc disk", 00:20:28.613 "block_size": 512, 00:20:28.613 "num_blocks": 65536, 00:20:28.613 "uuid": "aa941b85-da99-4820-a3d2-13b6a77a4bb1", 00:20:28.613 "assigned_rate_limits": { 00:20:28.613 "rw_ios_per_sec": 0, 00:20:28.613 "rw_mbytes_per_sec": 0, 00:20:28.613 "r_mbytes_per_sec": 0, 00:20:28.614 "w_mbytes_per_sec": 0 00:20:28.614 }, 00:20:28.614 "claimed": true, 00:20:28.614 "claim_type": "exclusive_write", 00:20:28.614 "zoned": false, 00:20:28.614 "supported_io_types": { 00:20:28.614 "read": true, 00:20:28.614 "write": true, 00:20:28.614 "unmap": true, 00:20:28.614 "flush": true, 00:20:28.614 "reset": true, 00:20:28.614 "nvme_admin": false, 00:20:28.614 "nvme_io": false, 00:20:28.614 "nvme_io_md": false, 00:20:28.614 "write_zeroes": true, 00:20:28.614 "zcopy": true, 00:20:28.614 "get_zone_info": false, 00:20:28.614 "zone_management": false, 00:20:28.614 "zone_append": false, 00:20:28.614 "compare": false, 00:20:28.614 "compare_and_write": false, 00:20:28.614 "abort": true, 00:20:28.614 "seek_hole": false, 00:20:28.614 "seek_data": false, 00:20:28.614 "copy": true, 00:20:28.614 "nvme_iov_md": false 00:20:28.614 }, 00:20:28.614 "memory_domains": [ 00:20:28.614 { 00:20:28.614 "dma_device_id": "system", 00:20:28.614 "dma_device_type": 1 00:20:28.614 }, 00:20:28.614 { 00:20:28.614 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:28.614 "dma_device_type": 2 00:20:28.614 } 00:20:28.614 ], 00:20:28.614 "driver_specific": {} 00:20:28.614 } 00:20:28.614 ] 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:28.614 "name": "Existed_Raid", 00:20:28.614 "uuid": "828f486c-5b7b-415a-9c26-a4f16a0fe496", 00:20:28.614 "strip_size_kb": 64, 00:20:28.614 "state": "online", 00:20:28.614 "raid_level": "concat", 00:20:28.614 "superblock": true, 00:20:28.614 "num_base_bdevs": 2, 00:20:28.614 "num_base_bdevs_discovered": 2, 00:20:28.614 "num_base_bdevs_operational": 2, 00:20:28.614 "base_bdevs_list": [ 00:20:28.614 { 00:20:28.614 "name": "BaseBdev1", 00:20:28.614 "uuid": "165a0c04-851a-4ae4-aafb-2ae1fe2a8749", 00:20:28.614 "is_configured": true, 00:20:28.614 "data_offset": 2048, 00:20:28.614 "data_size": 63488 00:20:28.614 }, 00:20:28.614 { 00:20:28.614 "name": "BaseBdev2", 00:20:28.614 "uuid": "aa941b85-da99-4820-a3d2-13b6a77a4bb1", 00:20:28.614 "is_configured": true, 00:20:28.614 "data_offset": 2048, 00:20:28.614 "data_size": 63488 00:20:28.614 } 00:20:28.614 ] 00:20:28.614 }' 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:28.614 20:19:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:29.264 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:20:29.264 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:20:29.264 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:20:29.264 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:20:29.264 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:20:29.264 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:20:29.264 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:20:29.264 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:20:29.264 20:19:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:29.264 20:19:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:29.264 [2024-10-01 20:19:24.212281] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:29.264 20:19:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:29.264 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:20:29.264 "name": "Existed_Raid", 00:20:29.264 "aliases": [ 00:20:29.264 "828f486c-5b7b-415a-9c26-a4f16a0fe496" 00:20:29.264 ], 00:20:29.264 "product_name": "Raid Volume", 00:20:29.264 "block_size": 512, 00:20:29.264 "num_blocks": 126976, 00:20:29.264 "uuid": "828f486c-5b7b-415a-9c26-a4f16a0fe496", 00:20:29.264 "assigned_rate_limits": { 00:20:29.264 "rw_ios_per_sec": 0, 00:20:29.264 "rw_mbytes_per_sec": 0, 00:20:29.264 "r_mbytes_per_sec": 0, 00:20:29.264 "w_mbytes_per_sec": 0 00:20:29.264 }, 00:20:29.264 "claimed": false, 00:20:29.264 "zoned": false, 00:20:29.264 "supported_io_types": { 00:20:29.264 "read": true, 00:20:29.264 "write": true, 00:20:29.264 "unmap": true, 00:20:29.264 "flush": true, 00:20:29.264 "reset": true, 00:20:29.264 "nvme_admin": false, 00:20:29.264 "nvme_io": false, 00:20:29.264 "nvme_io_md": false, 00:20:29.264 "write_zeroes": true, 00:20:29.264 "zcopy": false, 00:20:29.264 "get_zone_info": false, 00:20:29.264 "zone_management": false, 00:20:29.264 "zone_append": false, 00:20:29.264 "compare": false, 00:20:29.264 "compare_and_write": false, 00:20:29.264 "abort": false, 00:20:29.264 "seek_hole": false, 00:20:29.264 "seek_data": false, 00:20:29.264 "copy": false, 00:20:29.264 "nvme_iov_md": false 00:20:29.264 }, 00:20:29.264 "memory_domains": [ 00:20:29.264 { 00:20:29.264 "dma_device_id": "system", 00:20:29.264 "dma_device_type": 1 00:20:29.264 }, 00:20:29.264 { 00:20:29.264 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:29.264 "dma_device_type": 2 00:20:29.264 }, 00:20:29.264 { 00:20:29.264 "dma_device_id": "system", 00:20:29.264 "dma_device_type": 1 00:20:29.264 }, 00:20:29.264 { 00:20:29.264 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:29.264 "dma_device_type": 2 00:20:29.264 } 00:20:29.264 ], 00:20:29.264 "driver_specific": { 00:20:29.264 "raid": { 00:20:29.264 "uuid": "828f486c-5b7b-415a-9c26-a4f16a0fe496", 00:20:29.264 "strip_size_kb": 64, 00:20:29.264 "state": "online", 00:20:29.264 "raid_level": "concat", 00:20:29.264 "superblock": true, 00:20:29.264 "num_base_bdevs": 2, 00:20:29.264 "num_base_bdevs_discovered": 2, 00:20:29.264 "num_base_bdevs_operational": 2, 00:20:29.264 "base_bdevs_list": [ 00:20:29.264 { 00:20:29.264 "name": "BaseBdev1", 00:20:29.264 "uuid": "165a0c04-851a-4ae4-aafb-2ae1fe2a8749", 00:20:29.264 "is_configured": true, 00:20:29.264 "data_offset": 2048, 00:20:29.264 "data_size": 63488 00:20:29.264 }, 00:20:29.264 { 00:20:29.264 "name": "BaseBdev2", 00:20:29.264 "uuid": "aa941b85-da99-4820-a3d2-13b6a77a4bb1", 00:20:29.264 "is_configured": true, 00:20:29.264 "data_offset": 2048, 00:20:29.264 "data_size": 63488 00:20:29.264 } 00:20:29.264 ] 00:20:29.264 } 00:20:29.264 } 00:20:29.264 }' 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:20:29.265 BaseBdev2' 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:29.265 20:19:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:29.265 [2024-10-01 20:19:24.464042] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:20:29.265 [2024-10-01 20:19:24.464090] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:20:29.265 [2024-10-01 20:19:24.464163] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:29.529 "name": "Existed_Raid", 00:20:29.529 "uuid": "828f486c-5b7b-415a-9c26-a4f16a0fe496", 00:20:29.529 "strip_size_kb": 64, 00:20:29.529 "state": "offline", 00:20:29.529 "raid_level": "concat", 00:20:29.529 "superblock": true, 00:20:29.529 "num_base_bdevs": 2, 00:20:29.529 "num_base_bdevs_discovered": 1, 00:20:29.529 "num_base_bdevs_operational": 1, 00:20:29.529 "base_bdevs_list": [ 00:20:29.529 { 00:20:29.529 "name": null, 00:20:29.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:29.529 "is_configured": false, 00:20:29.529 "data_offset": 0, 00:20:29.529 "data_size": 63488 00:20:29.529 }, 00:20:29.529 { 00:20:29.529 "name": "BaseBdev2", 00:20:29.529 "uuid": "aa941b85-da99-4820-a3d2-13b6a77a4bb1", 00:20:29.529 "is_configured": true, 00:20:29.529 "data_offset": 2048, 00:20:29.529 "data_size": 63488 00:20:29.529 } 00:20:29.529 ] 00:20:29.529 }' 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:29.529 20:19:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:30.115 [2024-10-01 20:19:25.152781] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:20:30.115 [2024-10-01 20:19:25.152888] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 62340 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 62340 ']' 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 62340 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 62340 00:20:30.115 killing process with pid 62340 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 62340' 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 62340 00:20:30.115 [2024-10-01 20:19:25.329211] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:20:30.115 20:19:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 62340 00:20:30.115 [2024-10-01 20:19:25.344703] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:20:32.026 20:19:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:20:32.026 00:20:32.026 real 0m6.362s 00:20:32.026 user 0m9.076s 00:20:32.026 sys 0m0.922s 00:20:32.026 ************************************ 00:20:32.026 END TEST raid_state_function_test_sb 00:20:32.026 ************************************ 00:20:32.026 20:19:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:20:32.026 20:19:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:32.026 20:19:27 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 2 00:20:32.026 20:19:27 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:20:32.026 20:19:27 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:20:32.026 20:19:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:20:32.026 ************************************ 00:20:32.026 START TEST raid_superblock_test 00:20:32.026 ************************************ 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test concat 2 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=62603 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 62603 00:20:32.026 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 62603 ']' 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:32.026 20:19:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:32.285 [2024-10-01 20:19:27.351367] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:20:32.285 [2024-10-01 20:19:27.351604] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid62603 ] 00:20:32.543 [2024-10-01 20:19:27.538497] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:32.543 [2024-10-01 20:19:27.772655] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:20:32.801 [2024-10-01 20:19:27.977712] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:32.801 [2024-10-01 20:19:27.977762] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:33.369 malloc1 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:33.369 [2024-10-01 20:19:28.434086] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:20:33.369 [2024-10-01 20:19:28.434193] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:33.369 [2024-10-01 20:19:28.434227] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:20:33.369 [2024-10-01 20:19:28.434248] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:33.369 [2024-10-01 20:19:28.437192] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:33.369 [2024-10-01 20:19:28.437256] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:20:33.369 pt1 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:33.369 malloc2 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:33.369 [2024-10-01 20:19:28.485253] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:20:33.369 [2024-10-01 20:19:28.485527] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:33.369 [2024-10-01 20:19:28.485580] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:20:33.369 [2024-10-01 20:19:28.485599] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:33.369 [2024-10-01 20:19:28.488631] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:33.369 [2024-10-01 20:19:28.488860] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:20:33.369 pt2 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:33.369 [2024-10-01 20:19:28.493555] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:20:33.369 [2024-10-01 20:19:28.496251] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:20:33.369 [2024-10-01 20:19:28.496450] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:20:33.369 [2024-10-01 20:19:28.496468] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:20:33.369 [2024-10-01 20:19:28.496794] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:20:33.369 [2024-10-01 20:19:28.497028] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:20:33.369 [2024-10-01 20:19:28.497049] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:20:33.369 [2024-10-01 20:19:28.497227] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:33.369 "name": "raid_bdev1", 00:20:33.369 "uuid": "b2b016d2-18d0-42cd-a06a-3ffc57479a3d", 00:20:33.369 "strip_size_kb": 64, 00:20:33.369 "state": "online", 00:20:33.369 "raid_level": "concat", 00:20:33.369 "superblock": true, 00:20:33.369 "num_base_bdevs": 2, 00:20:33.369 "num_base_bdevs_discovered": 2, 00:20:33.369 "num_base_bdevs_operational": 2, 00:20:33.369 "base_bdevs_list": [ 00:20:33.369 { 00:20:33.369 "name": "pt1", 00:20:33.369 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:33.369 "is_configured": true, 00:20:33.369 "data_offset": 2048, 00:20:33.369 "data_size": 63488 00:20:33.369 }, 00:20:33.369 { 00:20:33.369 "name": "pt2", 00:20:33.369 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:33.369 "is_configured": true, 00:20:33.369 "data_offset": 2048, 00:20:33.369 "data_size": 63488 00:20:33.369 } 00:20:33.369 ] 00:20:33.369 }' 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:33.369 20:19:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:33.937 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:20:33.937 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:20:33.937 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:20:33.937 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:20:33.937 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:20:33.937 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:20:33.937 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:20:33.937 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:20:33.937 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:33.937 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:33.937 [2024-10-01 20:19:29.030051] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:33.937 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:33.937 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:20:33.937 "name": "raid_bdev1", 00:20:33.937 "aliases": [ 00:20:33.937 "b2b016d2-18d0-42cd-a06a-3ffc57479a3d" 00:20:33.938 ], 00:20:33.938 "product_name": "Raid Volume", 00:20:33.938 "block_size": 512, 00:20:33.938 "num_blocks": 126976, 00:20:33.938 "uuid": "b2b016d2-18d0-42cd-a06a-3ffc57479a3d", 00:20:33.938 "assigned_rate_limits": { 00:20:33.938 "rw_ios_per_sec": 0, 00:20:33.938 "rw_mbytes_per_sec": 0, 00:20:33.938 "r_mbytes_per_sec": 0, 00:20:33.938 "w_mbytes_per_sec": 0 00:20:33.938 }, 00:20:33.938 "claimed": false, 00:20:33.938 "zoned": false, 00:20:33.938 "supported_io_types": { 00:20:33.938 "read": true, 00:20:33.938 "write": true, 00:20:33.938 "unmap": true, 00:20:33.938 "flush": true, 00:20:33.938 "reset": true, 00:20:33.938 "nvme_admin": false, 00:20:33.938 "nvme_io": false, 00:20:33.938 "nvme_io_md": false, 00:20:33.938 "write_zeroes": true, 00:20:33.938 "zcopy": false, 00:20:33.938 "get_zone_info": false, 00:20:33.938 "zone_management": false, 00:20:33.938 "zone_append": false, 00:20:33.938 "compare": false, 00:20:33.938 "compare_and_write": false, 00:20:33.938 "abort": false, 00:20:33.938 "seek_hole": false, 00:20:33.938 "seek_data": false, 00:20:33.938 "copy": false, 00:20:33.938 "nvme_iov_md": false 00:20:33.938 }, 00:20:33.938 "memory_domains": [ 00:20:33.938 { 00:20:33.938 "dma_device_id": "system", 00:20:33.938 "dma_device_type": 1 00:20:33.938 }, 00:20:33.938 { 00:20:33.938 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:33.938 "dma_device_type": 2 00:20:33.938 }, 00:20:33.938 { 00:20:33.938 "dma_device_id": "system", 00:20:33.938 "dma_device_type": 1 00:20:33.938 }, 00:20:33.938 { 00:20:33.938 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:33.938 "dma_device_type": 2 00:20:33.938 } 00:20:33.938 ], 00:20:33.938 "driver_specific": { 00:20:33.938 "raid": { 00:20:33.938 "uuid": "b2b016d2-18d0-42cd-a06a-3ffc57479a3d", 00:20:33.938 "strip_size_kb": 64, 00:20:33.938 "state": "online", 00:20:33.938 "raid_level": "concat", 00:20:33.938 "superblock": true, 00:20:33.938 "num_base_bdevs": 2, 00:20:33.938 "num_base_bdevs_discovered": 2, 00:20:33.938 "num_base_bdevs_operational": 2, 00:20:33.938 "base_bdevs_list": [ 00:20:33.938 { 00:20:33.938 "name": "pt1", 00:20:33.938 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:33.938 "is_configured": true, 00:20:33.938 "data_offset": 2048, 00:20:33.938 "data_size": 63488 00:20:33.938 }, 00:20:33.938 { 00:20:33.938 "name": "pt2", 00:20:33.938 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:33.938 "is_configured": true, 00:20:33.938 "data_offset": 2048, 00:20:33.938 "data_size": 63488 00:20:33.938 } 00:20:33.938 ] 00:20:33.938 } 00:20:33.938 } 00:20:33.938 }' 00:20:33.938 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:20:33.938 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:20:33.938 pt2' 00:20:33.938 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:34.198 [2024-10-01 20:19:29.314090] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=b2b016d2-18d0-42cd-a06a-3ffc57479a3d 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z b2b016d2-18d0-42cd-a06a-3ffc57479a3d ']' 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:34.198 [2024-10-01 20:19:29.361737] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:20:34.198 [2024-10-01 20:19:29.361794] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:20:34.198 [2024-10-01 20:19:29.361905] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:34.198 [2024-10-01 20:19:29.361975] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:20:34.198 [2024-10-01 20:19:29.362000] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:34.198 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:34.458 [2024-10-01 20:19:29.505870] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:20:34.458 [2024-10-01 20:19:29.508627] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:20:34.458 [2024-10-01 20:19:29.508728] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:20:34.458 [2024-10-01 20:19:29.508847] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:20:34.458 [2024-10-01 20:19:29.508891] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:20:34.458 [2024-10-01 20:19:29.508910] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:20:34.458 request: 00:20:34.458 { 00:20:34.458 "name": "raid_bdev1", 00:20:34.458 "raid_level": "concat", 00:20:34.458 "base_bdevs": [ 00:20:34.458 "malloc1", 00:20:34.458 "malloc2" 00:20:34.458 ], 00:20:34.458 "strip_size_kb": 64, 00:20:34.458 "superblock": false, 00:20:34.458 "method": "bdev_raid_create", 00:20:34.458 "req_id": 1 00:20:34.458 } 00:20:34.458 Got JSON-RPC error response 00:20:34.458 response: 00:20:34.458 { 00:20:34.458 "code": -17, 00:20:34.458 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:20:34.458 } 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:20:34.458 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:34.459 [2024-10-01 20:19:29.565865] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:20:34.459 [2024-10-01 20:19:29.566099] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:34.459 [2024-10-01 20:19:29.566177] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:20:34.459 [2024-10-01 20:19:29.566422] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:34.459 [2024-10-01 20:19:29.569709] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:34.459 [2024-10-01 20:19:29.569927] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:20:34.459 [2024-10-01 20:19:29.570150] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:20:34.459 [2024-10-01 20:19:29.570349] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:20:34.459 pt1 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 2 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:34.459 "name": "raid_bdev1", 00:20:34.459 "uuid": "b2b016d2-18d0-42cd-a06a-3ffc57479a3d", 00:20:34.459 "strip_size_kb": 64, 00:20:34.459 "state": "configuring", 00:20:34.459 "raid_level": "concat", 00:20:34.459 "superblock": true, 00:20:34.459 "num_base_bdevs": 2, 00:20:34.459 "num_base_bdevs_discovered": 1, 00:20:34.459 "num_base_bdevs_operational": 2, 00:20:34.459 "base_bdevs_list": [ 00:20:34.459 { 00:20:34.459 "name": "pt1", 00:20:34.459 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:34.459 "is_configured": true, 00:20:34.459 "data_offset": 2048, 00:20:34.459 "data_size": 63488 00:20:34.459 }, 00:20:34.459 { 00:20:34.459 "name": null, 00:20:34.459 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:34.459 "is_configured": false, 00:20:34.459 "data_offset": 2048, 00:20:34.459 "data_size": 63488 00:20:34.459 } 00:20:34.459 ] 00:20:34.459 }' 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:34.459 20:19:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:35.109 [2024-10-01 20:19:30.074431] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:20:35.109 [2024-10-01 20:19:30.074540] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:35.109 [2024-10-01 20:19:30.074575] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:20:35.109 [2024-10-01 20:19:30.074596] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:35.109 [2024-10-01 20:19:30.075284] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:35.109 [2024-10-01 20:19:30.075321] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:20:35.109 [2024-10-01 20:19:30.075420] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:20:35.109 [2024-10-01 20:19:30.075489] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:20:35.109 [2024-10-01 20:19:30.075628] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:20:35.109 [2024-10-01 20:19:30.075648] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:20:35.109 [2024-10-01 20:19:30.076012] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:20:35.109 [2024-10-01 20:19:30.076223] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:20:35.109 [2024-10-01 20:19:30.076239] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:20:35.109 [2024-10-01 20:19:30.076408] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:35.109 pt2 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:35.109 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:35.109 "name": "raid_bdev1", 00:20:35.109 "uuid": "b2b016d2-18d0-42cd-a06a-3ffc57479a3d", 00:20:35.109 "strip_size_kb": 64, 00:20:35.109 "state": "online", 00:20:35.109 "raid_level": "concat", 00:20:35.109 "superblock": true, 00:20:35.109 "num_base_bdevs": 2, 00:20:35.109 "num_base_bdevs_discovered": 2, 00:20:35.109 "num_base_bdevs_operational": 2, 00:20:35.109 "base_bdevs_list": [ 00:20:35.109 { 00:20:35.109 "name": "pt1", 00:20:35.109 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:35.109 "is_configured": true, 00:20:35.109 "data_offset": 2048, 00:20:35.109 "data_size": 63488 00:20:35.109 }, 00:20:35.109 { 00:20:35.109 "name": "pt2", 00:20:35.109 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:35.109 "is_configured": true, 00:20:35.109 "data_offset": 2048, 00:20:35.110 "data_size": 63488 00:20:35.110 } 00:20:35.110 ] 00:20:35.110 }' 00:20:35.110 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:35.110 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:35.368 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:20:35.368 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:20:35.368 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:20:35.368 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:20:35.368 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:20:35.368 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:20:35.368 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:20:35.368 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:20:35.368 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:35.368 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:35.368 [2024-10-01 20:19:30.602983] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:35.627 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:35.627 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:20:35.627 "name": "raid_bdev1", 00:20:35.627 "aliases": [ 00:20:35.627 "b2b016d2-18d0-42cd-a06a-3ffc57479a3d" 00:20:35.627 ], 00:20:35.627 "product_name": "Raid Volume", 00:20:35.627 "block_size": 512, 00:20:35.627 "num_blocks": 126976, 00:20:35.627 "uuid": "b2b016d2-18d0-42cd-a06a-3ffc57479a3d", 00:20:35.627 "assigned_rate_limits": { 00:20:35.627 "rw_ios_per_sec": 0, 00:20:35.627 "rw_mbytes_per_sec": 0, 00:20:35.627 "r_mbytes_per_sec": 0, 00:20:35.627 "w_mbytes_per_sec": 0 00:20:35.627 }, 00:20:35.627 "claimed": false, 00:20:35.627 "zoned": false, 00:20:35.627 "supported_io_types": { 00:20:35.627 "read": true, 00:20:35.627 "write": true, 00:20:35.627 "unmap": true, 00:20:35.627 "flush": true, 00:20:35.627 "reset": true, 00:20:35.627 "nvme_admin": false, 00:20:35.627 "nvme_io": false, 00:20:35.627 "nvme_io_md": false, 00:20:35.627 "write_zeroes": true, 00:20:35.627 "zcopy": false, 00:20:35.627 "get_zone_info": false, 00:20:35.627 "zone_management": false, 00:20:35.627 "zone_append": false, 00:20:35.627 "compare": false, 00:20:35.627 "compare_and_write": false, 00:20:35.627 "abort": false, 00:20:35.627 "seek_hole": false, 00:20:35.627 "seek_data": false, 00:20:35.627 "copy": false, 00:20:35.627 "nvme_iov_md": false 00:20:35.628 }, 00:20:35.628 "memory_domains": [ 00:20:35.628 { 00:20:35.628 "dma_device_id": "system", 00:20:35.628 "dma_device_type": 1 00:20:35.628 }, 00:20:35.628 { 00:20:35.628 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:35.628 "dma_device_type": 2 00:20:35.628 }, 00:20:35.628 { 00:20:35.628 "dma_device_id": "system", 00:20:35.628 "dma_device_type": 1 00:20:35.628 }, 00:20:35.628 { 00:20:35.628 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:35.628 "dma_device_type": 2 00:20:35.628 } 00:20:35.628 ], 00:20:35.628 "driver_specific": { 00:20:35.628 "raid": { 00:20:35.628 "uuid": "b2b016d2-18d0-42cd-a06a-3ffc57479a3d", 00:20:35.628 "strip_size_kb": 64, 00:20:35.628 "state": "online", 00:20:35.628 "raid_level": "concat", 00:20:35.628 "superblock": true, 00:20:35.628 "num_base_bdevs": 2, 00:20:35.628 "num_base_bdevs_discovered": 2, 00:20:35.628 "num_base_bdevs_operational": 2, 00:20:35.628 "base_bdevs_list": [ 00:20:35.628 { 00:20:35.628 "name": "pt1", 00:20:35.628 "uuid": "00000000-0000-0000-0000-000000000001", 00:20:35.628 "is_configured": true, 00:20:35.628 "data_offset": 2048, 00:20:35.628 "data_size": 63488 00:20:35.628 }, 00:20:35.628 { 00:20:35.628 "name": "pt2", 00:20:35.628 "uuid": "00000000-0000-0000-0000-000000000002", 00:20:35.628 "is_configured": true, 00:20:35.628 "data_offset": 2048, 00:20:35.628 "data_size": 63488 00:20:35.628 } 00:20:35.628 ] 00:20:35.628 } 00:20:35.628 } 00:20:35.628 }' 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:20:35.628 pt2' 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:35.628 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:35.886 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:35.886 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:35.886 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:20:35.886 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:35.886 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:35.886 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:20:35.886 [2024-10-01 20:19:30.910968] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:35.886 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:35.886 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' b2b016d2-18d0-42cd-a06a-3ffc57479a3d '!=' b2b016d2-18d0-42cd-a06a-3ffc57479a3d ']' 00:20:35.887 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:20:35.887 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:20:35.887 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:20:35.887 20:19:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 62603 00:20:35.887 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 62603 ']' 00:20:35.887 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 62603 00:20:35.887 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:20:35.887 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:20:35.887 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 62603 00:20:35.887 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:20:35.887 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:20:35.887 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 62603' 00:20:35.887 killing process with pid 62603 00:20:35.887 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 62603 00:20:35.887 [2024-10-01 20:19:30.999813] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:20:35.887 20:19:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 62603 00:20:35.887 [2024-10-01 20:19:30.999984] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:35.887 [2024-10-01 20:19:31.000058] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:20:35.887 [2024-10-01 20:19:31.000086] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:20:36.145 [2024-10-01 20:19:31.178450] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:20:38.049 20:19:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:20:38.049 00:20:38.049 real 0m5.789s 00:20:38.049 user 0m7.967s 00:20:38.049 sys 0m0.883s 00:20:38.049 20:19:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:20:38.049 ************************************ 00:20:38.049 END TEST raid_superblock_test 00:20:38.049 ************************************ 00:20:38.049 20:19:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:20:38.049 20:19:33 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 2 read 00:20:38.049 20:19:33 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:20:38.049 20:19:33 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:20:38.049 20:19:33 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:20:38.049 ************************************ 00:20:38.049 START TEST raid_read_error_test 00:20:38.049 ************************************ 00:20:38.049 20:19:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 2 read 00:20:38.049 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:20:38.049 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:20:38.049 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:20:38.049 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:20:38.049 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:20:38.049 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:20:38.049 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:20:38.049 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:20:38.049 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:20:38.049 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.M93DhJqBSN 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=62826 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 62826 00:20:38.050 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 62826 ']' 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:38.050 20:19:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:38.050 [2024-10-01 20:19:33.198524] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:20:38.050 [2024-10-01 20:19:33.199007] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid62826 ] 00:20:38.308 [2024-10-01 20:19:33.390907] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:38.568 [2024-10-01 20:19:33.659314] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:20:38.827 [2024-10-01 20:19:33.869479] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:38.827 [2024-10-01 20:19:33.869575] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:39.087 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:39.087 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:20:39.087 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:20:39.087 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:20:39.087 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:39.087 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:39.087 BaseBdev1_malloc 00:20:39.087 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:39.087 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:20:39.087 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:39.087 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:39.346 true 00:20:39.346 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:39.346 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:20:39.346 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:39.346 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:39.346 [2024-10-01 20:19:34.355908] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:20:39.346 [2024-10-01 20:19:34.356152] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:39.346 [2024-10-01 20:19:34.356190] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:20:39.346 [2024-10-01 20:19:34.356213] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:39.346 [2024-10-01 20:19:34.359240] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:39.346 [2024-10-01 20:19:34.359292] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:20:39.346 BaseBdev1 00:20:39.346 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:39.346 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:20:39.346 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:20:39.346 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:39.346 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:39.346 BaseBdev2_malloc 00:20:39.346 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:39.346 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:20:39.346 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:39.346 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:39.346 true 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:39.347 [2024-10-01 20:19:34.425734] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:20:39.347 [2024-10-01 20:19:34.425841] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:39.347 [2024-10-01 20:19:34.425875] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:20:39.347 [2024-10-01 20:19:34.425895] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:39.347 [2024-10-01 20:19:34.428916] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:39.347 [2024-10-01 20:19:34.428999] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:20:39.347 BaseBdev2 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:39.347 [2024-10-01 20:19:34.437958] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:20:39.347 [2024-10-01 20:19:34.440806] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:20:39.347 [2024-10-01 20:19:34.441207] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:20:39.347 [2024-10-01 20:19:34.441351] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:20:39.347 [2024-10-01 20:19:34.441785] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:20:39.347 [2024-10-01 20:19:34.442135] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:20:39.347 [2024-10-01 20:19:34.442255] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:20:39.347 [2024-10-01 20:19:34.442642] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:39.347 "name": "raid_bdev1", 00:20:39.347 "uuid": "0cbb0597-f03e-4d16-8d85-655811d194ca", 00:20:39.347 "strip_size_kb": 64, 00:20:39.347 "state": "online", 00:20:39.347 "raid_level": "concat", 00:20:39.347 "superblock": true, 00:20:39.347 "num_base_bdevs": 2, 00:20:39.347 "num_base_bdevs_discovered": 2, 00:20:39.347 "num_base_bdevs_operational": 2, 00:20:39.347 "base_bdevs_list": [ 00:20:39.347 { 00:20:39.347 "name": "BaseBdev1", 00:20:39.347 "uuid": "878b127c-ad2d-5ba4-aaa0-0b11e60bafe4", 00:20:39.347 "is_configured": true, 00:20:39.347 "data_offset": 2048, 00:20:39.347 "data_size": 63488 00:20:39.347 }, 00:20:39.347 { 00:20:39.347 "name": "BaseBdev2", 00:20:39.347 "uuid": "12c17a92-530a-5cb7-b44c-02ed68b73099", 00:20:39.347 "is_configured": true, 00:20:39.347 "data_offset": 2048, 00:20:39.347 "data_size": 63488 00:20:39.347 } 00:20:39.347 ] 00:20:39.347 }' 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:39.347 20:19:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:39.915 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:20:39.915 20:19:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:20:39.915 [2024-10-01 20:19:35.068261] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:40.854 20:19:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:40.854 20:19:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:40.854 "name": "raid_bdev1", 00:20:40.854 "uuid": "0cbb0597-f03e-4d16-8d85-655811d194ca", 00:20:40.854 "strip_size_kb": 64, 00:20:40.854 "state": "online", 00:20:40.854 "raid_level": "concat", 00:20:40.854 "superblock": true, 00:20:40.854 "num_base_bdevs": 2, 00:20:40.854 "num_base_bdevs_discovered": 2, 00:20:40.854 "num_base_bdevs_operational": 2, 00:20:40.854 "base_bdevs_list": [ 00:20:40.854 { 00:20:40.854 "name": "BaseBdev1", 00:20:40.854 "uuid": "878b127c-ad2d-5ba4-aaa0-0b11e60bafe4", 00:20:40.854 "is_configured": true, 00:20:40.854 "data_offset": 2048, 00:20:40.854 "data_size": 63488 00:20:40.854 }, 00:20:40.854 { 00:20:40.854 "name": "BaseBdev2", 00:20:40.854 "uuid": "12c17a92-530a-5cb7-b44c-02ed68b73099", 00:20:40.854 "is_configured": true, 00:20:40.854 "data_offset": 2048, 00:20:40.854 "data_size": 63488 00:20:40.854 } 00:20:40.854 ] 00:20:40.854 }' 00:20:40.854 20:19:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:40.854 20:19:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:41.423 20:19:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:20:41.423 20:19:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:41.423 20:19:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:41.423 [2024-10-01 20:19:36.472050] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:20:41.423 [2024-10-01 20:19:36.472106] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:20:41.423 [2024-10-01 20:19:36.475638] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:41.423 [2024-10-01 20:19:36.475894] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:41.423 [2024-10-01 20:19:36.475964] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:20:41.423 [2024-10-01 20:19:36.475986] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:20:41.423 { 00:20:41.423 "results": [ 00:20:41.423 { 00:20:41.423 "job": "raid_bdev1", 00:20:41.423 "core_mask": "0x1", 00:20:41.423 "workload": "randrw", 00:20:41.423 "percentage": 50, 00:20:41.423 "status": "finished", 00:20:41.423 "queue_depth": 1, 00:20:41.423 "io_size": 131072, 00:20:41.423 "runtime": 1.400926, 00:20:41.423 "iops": 10390.984248989596, 00:20:41.423 "mibps": 1298.8730311236995, 00:20:41.423 "io_failed": 1, 00:20:41.423 "io_timeout": 0, 00:20:41.423 "avg_latency_us": 134.1341972548677, 00:20:41.423 "min_latency_us": 40.02909090909091, 00:20:41.423 "max_latency_us": 1891.6072727272726 00:20:41.423 } 00:20:41.423 ], 00:20:41.423 "core_count": 1 00:20:41.423 } 00:20:41.423 20:19:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:41.423 20:19:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 62826 00:20:41.423 20:19:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 62826 ']' 00:20:41.423 20:19:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 62826 00:20:41.423 20:19:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:20:41.424 20:19:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:20:41.424 20:19:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 62826 00:20:41.424 killing process with pid 62826 00:20:41.424 20:19:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:20:41.424 20:19:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:20:41.424 20:19:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 62826' 00:20:41.424 20:19:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 62826 00:20:41.424 [2024-10-01 20:19:36.515041] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:20:41.424 20:19:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 62826 00:20:41.424 [2024-10-01 20:19:36.642151] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:20:43.327 20:19:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.M93DhJqBSN 00:20:43.327 20:19:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:20:43.327 20:19:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:20:43.327 ************************************ 00:20:43.327 END TEST raid_read_error_test 00:20:43.327 ************************************ 00:20:43.327 20:19:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.71 00:20:43.327 20:19:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:20:43.327 20:19:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:20:43.327 20:19:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:20:43.327 20:19:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.71 != \0\.\0\0 ]] 00:20:43.327 00:20:43.327 real 0m5.441s 00:20:43.327 user 0m6.526s 00:20:43.327 sys 0m0.684s 00:20:43.327 20:19:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:20:43.327 20:19:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:43.327 20:19:38 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 2 write 00:20:43.327 20:19:38 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:20:43.327 20:19:38 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:20:43.327 20:19:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:20:43.327 ************************************ 00:20:43.327 START TEST raid_write_error_test 00:20:43.327 ************************************ 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 2 write 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:20:43.327 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:20:43.587 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.4bNKfe7w4L 00:20:43.587 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=62977 00:20:43.587 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:20:43.587 20:19:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 62977 00:20:43.587 20:19:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 62977 ']' 00:20:43.587 20:19:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:43.587 20:19:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:43.587 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:43.587 20:19:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:43.587 20:19:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:43.587 20:19:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:43.587 [2024-10-01 20:19:38.701120] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:20:43.587 [2024-10-01 20:19:38.701637] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid62977 ] 00:20:43.853 [2024-10-01 20:19:38.880825] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:44.115 [2024-10-01 20:19:39.124676] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:20:44.115 [2024-10-01 20:19:39.330867] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:44.115 [2024-10-01 20:19:39.330960] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:44.683 BaseBdev1_malloc 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:44.683 true 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:44.683 [2024-10-01 20:19:39.797718] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:20:44.683 [2024-10-01 20:19:39.797857] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:44.683 [2024-10-01 20:19:39.797887] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:20:44.683 [2024-10-01 20:19:39.797923] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:44.683 [2024-10-01 20:19:39.800955] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:44.683 [2024-10-01 20:19:39.801010] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:20:44.683 BaseBdev1 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:44.683 BaseBdev2_malloc 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:44.683 true 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:44.683 [2024-10-01 20:19:39.854074] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:20:44.683 [2024-10-01 20:19:39.854168] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:20:44.683 [2024-10-01 20:19:39.854198] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:20:44.683 [2024-10-01 20:19:39.854216] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:20:44.683 [2024-10-01 20:19:39.857376] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:20:44.683 [2024-10-01 20:19:39.857462] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:20:44.683 BaseBdev2 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:44.683 [2024-10-01 20:19:39.862213] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:20:44.683 [2024-10-01 20:19:39.864846] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:20:44.683 [2024-10-01 20:19:39.865134] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:20:44.683 [2024-10-01 20:19:39.865167] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:20:44.683 [2024-10-01 20:19:39.865494] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:20:44.683 [2024-10-01 20:19:39.865779] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:20:44.683 [2024-10-01 20:19:39.865797] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:20:44.683 [2024-10-01 20:19:39.866029] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:44.683 "name": "raid_bdev1", 00:20:44.683 "uuid": "2fdb9632-5012-4998-a36b-7a1b95334115", 00:20:44.683 "strip_size_kb": 64, 00:20:44.683 "state": "online", 00:20:44.683 "raid_level": "concat", 00:20:44.683 "superblock": true, 00:20:44.683 "num_base_bdevs": 2, 00:20:44.683 "num_base_bdevs_discovered": 2, 00:20:44.683 "num_base_bdevs_operational": 2, 00:20:44.683 "base_bdevs_list": [ 00:20:44.683 { 00:20:44.683 "name": "BaseBdev1", 00:20:44.683 "uuid": "50767aa1-c9c1-55eb-9169-a6193743d679", 00:20:44.683 "is_configured": true, 00:20:44.683 "data_offset": 2048, 00:20:44.683 "data_size": 63488 00:20:44.683 }, 00:20:44.683 { 00:20:44.683 "name": "BaseBdev2", 00:20:44.683 "uuid": "4f781c5f-ff84-530e-a283-875faf4c5d35", 00:20:44.683 "is_configured": true, 00:20:44.683 "data_offset": 2048, 00:20:44.683 "data_size": 63488 00:20:44.683 } 00:20:44.683 ] 00:20:44.683 }' 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:44.683 20:19:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:45.250 20:19:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:20:45.250 20:19:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:20:45.509 [2024-10-01 20:19:40.520092] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:20:46.443 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:20:46.443 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:46.443 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:46.443 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:46.443 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:20:46.443 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:20:46.443 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:20:46.443 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:20:46.443 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:20:46.443 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:46.443 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:20:46.444 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:20:46.444 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:46.444 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:46.444 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:46.444 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:46.444 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:46.444 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:46.444 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:20:46.444 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:46.444 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:46.444 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:46.444 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:46.444 "name": "raid_bdev1", 00:20:46.444 "uuid": "2fdb9632-5012-4998-a36b-7a1b95334115", 00:20:46.444 "strip_size_kb": 64, 00:20:46.444 "state": "online", 00:20:46.444 "raid_level": "concat", 00:20:46.444 "superblock": true, 00:20:46.444 "num_base_bdevs": 2, 00:20:46.444 "num_base_bdevs_discovered": 2, 00:20:46.444 "num_base_bdevs_operational": 2, 00:20:46.444 "base_bdevs_list": [ 00:20:46.444 { 00:20:46.444 "name": "BaseBdev1", 00:20:46.444 "uuid": "50767aa1-c9c1-55eb-9169-a6193743d679", 00:20:46.444 "is_configured": true, 00:20:46.444 "data_offset": 2048, 00:20:46.444 "data_size": 63488 00:20:46.444 }, 00:20:46.444 { 00:20:46.444 "name": "BaseBdev2", 00:20:46.444 "uuid": "4f781c5f-ff84-530e-a283-875faf4c5d35", 00:20:46.444 "is_configured": true, 00:20:46.444 "data_offset": 2048, 00:20:46.444 "data_size": 63488 00:20:46.444 } 00:20:46.444 ] 00:20:46.444 }' 00:20:46.444 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:46.444 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:46.702 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:20:46.702 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:46.702 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:46.702 [2024-10-01 20:19:41.947919] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:20:46.702 [2024-10-01 20:19:41.947966] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:20:46.702 [2024-10-01 20:19:41.951313] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:46.702 { 00:20:46.702 "results": [ 00:20:46.702 { 00:20:46.702 "job": "raid_bdev1", 00:20:46.702 "core_mask": "0x1", 00:20:46.702 "workload": "randrw", 00:20:46.702 "percentage": 50, 00:20:46.702 "status": "finished", 00:20:46.702 "queue_depth": 1, 00:20:46.702 "io_size": 131072, 00:20:46.702 "runtime": 1.42524, 00:20:46.702 "iops": 10441.750161376329, 00:20:46.702 "mibps": 1305.218770172041, 00:20:46.702 "io_failed": 1, 00:20:46.702 "io_timeout": 0, 00:20:46.702 "avg_latency_us": 133.538863743258, 00:20:46.702 "min_latency_us": 40.02909090909091, 00:20:46.702 "max_latency_us": 1884.16 00:20:46.702 } 00:20:46.702 ], 00:20:46.702 "core_count": 1 00:20:46.702 } 00:20:46.702 [2024-10-01 20:19:41.951525] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:46.702 [2024-10-01 20:19:41.951591] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:20:46.702 [2024-10-01 20:19:41.951613] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:20:46.965 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:46.965 20:19:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 62977 00:20:46.965 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 62977 ']' 00:20:46.965 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 62977 00:20:46.965 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:20:46.965 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:20:46.965 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 62977 00:20:46.965 killing process with pid 62977 00:20:46.965 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:20:46.965 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:20:46.965 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 62977' 00:20:46.965 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 62977 00:20:46.965 [2024-10-01 20:19:41.989497] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:20:46.965 20:19:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 62977 00:20:46.965 [2024-10-01 20:19:42.112751] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:20:48.876 20:19:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.4bNKfe7w4L 00:20:48.876 20:19:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:20:48.876 20:19:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:20:48.876 20:19:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.70 00:20:48.876 20:19:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:20:48.876 ************************************ 00:20:48.876 END TEST raid_write_error_test 00:20:48.876 ************************************ 00:20:48.876 20:19:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:20:48.876 20:19:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:20:48.876 20:19:43 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.70 != \0\.\0\0 ]] 00:20:48.876 00:20:48.876 real 0m5.416s 00:20:48.876 user 0m6.432s 00:20:48.876 sys 0m0.717s 00:20:48.876 20:19:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:20:48.876 20:19:43 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:20:48.876 20:19:44 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:20:48.877 20:19:44 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 2 false 00:20:48.877 20:19:44 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:20:48.877 20:19:44 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:20:48.877 20:19:44 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:20:48.877 ************************************ 00:20:48.877 START TEST raid_state_function_test 00:20:48.877 ************************************ 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 false 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=63126 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 63126' 00:20:48.877 Process raid pid: 63126 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 63126 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 63126 ']' 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:48.877 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:48.877 20:19:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:49.135 [2024-10-01 20:19:44.166765] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:20:49.135 [2024-10-01 20:19:44.167234] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:49.135 [2024-10-01 20:19:44.350455] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:49.700 [2024-10-01 20:19:44.648038] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:20:49.700 [2024-10-01 20:19:44.862384] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:49.700 [2024-10-01 20:19:44.862445] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:50.265 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:50.265 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:20:50.265 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:20:50.265 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:50.265 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:50.265 [2024-10-01 20:19:45.276021] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:20:50.265 [2024-10-01 20:19:45.276305] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:20:50.265 [2024-10-01 20:19:45.276354] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:20:50.265 [2024-10-01 20:19:45.276374] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:20:50.265 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:50.265 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:20:50.265 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:50.265 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:20:50.265 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:20:50.265 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:20:50.265 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:50.265 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:50.265 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:50.265 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:50.266 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:50.266 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:50.266 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:50.266 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:50.266 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:50.266 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:50.266 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:50.266 "name": "Existed_Raid", 00:20:50.266 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:50.266 "strip_size_kb": 0, 00:20:50.266 "state": "configuring", 00:20:50.266 "raid_level": "raid1", 00:20:50.266 "superblock": false, 00:20:50.266 "num_base_bdevs": 2, 00:20:50.266 "num_base_bdevs_discovered": 0, 00:20:50.266 "num_base_bdevs_operational": 2, 00:20:50.266 "base_bdevs_list": [ 00:20:50.266 { 00:20:50.266 "name": "BaseBdev1", 00:20:50.266 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:50.266 "is_configured": false, 00:20:50.266 "data_offset": 0, 00:20:50.266 "data_size": 0 00:20:50.266 }, 00:20:50.266 { 00:20:50.266 "name": "BaseBdev2", 00:20:50.266 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:50.266 "is_configured": false, 00:20:50.266 "data_offset": 0, 00:20:50.266 "data_size": 0 00:20:50.266 } 00:20:50.266 ] 00:20:50.266 }' 00:20:50.266 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:50.266 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:50.832 [2024-10-01 20:19:45.788134] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:20:50.832 [2024-10-01 20:19:45.788351] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:50.832 [2024-10-01 20:19:45.796054] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:20:50.832 [2024-10-01 20:19:45.796153] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:20:50.832 [2024-10-01 20:19:45.796168] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:20:50.832 [2024-10-01 20:19:45.796187] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:50.832 [2024-10-01 20:19:45.840570] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:20:50.832 BaseBdev1 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:20:50.832 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:50.833 [ 00:20:50.833 { 00:20:50.833 "name": "BaseBdev1", 00:20:50.833 "aliases": [ 00:20:50.833 "5b53a3c9-5e50-474d-a76c-293c38332991" 00:20:50.833 ], 00:20:50.833 "product_name": "Malloc disk", 00:20:50.833 "block_size": 512, 00:20:50.833 "num_blocks": 65536, 00:20:50.833 "uuid": "5b53a3c9-5e50-474d-a76c-293c38332991", 00:20:50.833 "assigned_rate_limits": { 00:20:50.833 "rw_ios_per_sec": 0, 00:20:50.833 "rw_mbytes_per_sec": 0, 00:20:50.833 "r_mbytes_per_sec": 0, 00:20:50.833 "w_mbytes_per_sec": 0 00:20:50.833 }, 00:20:50.833 "claimed": true, 00:20:50.833 "claim_type": "exclusive_write", 00:20:50.833 "zoned": false, 00:20:50.833 "supported_io_types": { 00:20:50.833 "read": true, 00:20:50.833 "write": true, 00:20:50.833 "unmap": true, 00:20:50.833 "flush": true, 00:20:50.833 "reset": true, 00:20:50.833 "nvme_admin": false, 00:20:50.833 "nvme_io": false, 00:20:50.833 "nvme_io_md": false, 00:20:50.833 "write_zeroes": true, 00:20:50.833 "zcopy": true, 00:20:50.833 "get_zone_info": false, 00:20:50.833 "zone_management": false, 00:20:50.833 "zone_append": false, 00:20:50.833 "compare": false, 00:20:50.833 "compare_and_write": false, 00:20:50.833 "abort": true, 00:20:50.833 "seek_hole": false, 00:20:50.833 "seek_data": false, 00:20:50.833 "copy": true, 00:20:50.833 "nvme_iov_md": false 00:20:50.833 }, 00:20:50.833 "memory_domains": [ 00:20:50.833 { 00:20:50.833 "dma_device_id": "system", 00:20:50.833 "dma_device_type": 1 00:20:50.833 }, 00:20:50.833 { 00:20:50.833 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:50.833 "dma_device_type": 2 00:20:50.833 } 00:20:50.833 ], 00:20:50.833 "driver_specific": {} 00:20:50.833 } 00:20:50.833 ] 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:50.833 "name": "Existed_Raid", 00:20:50.833 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:50.833 "strip_size_kb": 0, 00:20:50.833 "state": "configuring", 00:20:50.833 "raid_level": "raid1", 00:20:50.833 "superblock": false, 00:20:50.833 "num_base_bdevs": 2, 00:20:50.833 "num_base_bdevs_discovered": 1, 00:20:50.833 "num_base_bdevs_operational": 2, 00:20:50.833 "base_bdevs_list": [ 00:20:50.833 { 00:20:50.833 "name": "BaseBdev1", 00:20:50.833 "uuid": "5b53a3c9-5e50-474d-a76c-293c38332991", 00:20:50.833 "is_configured": true, 00:20:50.833 "data_offset": 0, 00:20:50.833 "data_size": 65536 00:20:50.833 }, 00:20:50.833 { 00:20:50.833 "name": "BaseBdev2", 00:20:50.833 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:50.833 "is_configured": false, 00:20:50.833 "data_offset": 0, 00:20:50.833 "data_size": 0 00:20:50.833 } 00:20:50.833 ] 00:20:50.833 }' 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:50.833 20:19:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:51.400 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:51.401 [2024-10-01 20:19:46.384843] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:20:51.401 [2024-10-01 20:19:46.384911] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:51.401 [2024-10-01 20:19:46.392944] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:20:51.401 [2024-10-01 20:19:46.395988] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:20:51.401 [2024-10-01 20:19:46.396060] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:51.401 "name": "Existed_Raid", 00:20:51.401 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:51.401 "strip_size_kb": 0, 00:20:51.401 "state": "configuring", 00:20:51.401 "raid_level": "raid1", 00:20:51.401 "superblock": false, 00:20:51.401 "num_base_bdevs": 2, 00:20:51.401 "num_base_bdevs_discovered": 1, 00:20:51.401 "num_base_bdevs_operational": 2, 00:20:51.401 "base_bdevs_list": [ 00:20:51.401 { 00:20:51.401 "name": "BaseBdev1", 00:20:51.401 "uuid": "5b53a3c9-5e50-474d-a76c-293c38332991", 00:20:51.401 "is_configured": true, 00:20:51.401 "data_offset": 0, 00:20:51.401 "data_size": 65536 00:20:51.401 }, 00:20:51.401 { 00:20:51.401 "name": "BaseBdev2", 00:20:51.401 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:51.401 "is_configured": false, 00:20:51.401 "data_offset": 0, 00:20:51.401 "data_size": 0 00:20:51.401 } 00:20:51.401 ] 00:20:51.401 }' 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:51.401 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:51.968 [2024-10-01 20:19:46.963906] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:20:51.968 [2024-10-01 20:19:46.964169] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:20:51.968 [2024-10-01 20:19:46.964200] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:20:51.968 [2024-10-01 20:19:46.964575] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:20:51.968 [2024-10-01 20:19:46.964834] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:20:51.968 [2024-10-01 20:19:46.964863] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:20:51.968 [2024-10-01 20:19:46.965275] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:51.968 BaseBdev2 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:51.968 [ 00:20:51.968 { 00:20:51.968 "name": "BaseBdev2", 00:20:51.968 "aliases": [ 00:20:51.968 "6d936500-88a8-4bb2-9901-c2f5866c6f8d" 00:20:51.968 ], 00:20:51.968 "product_name": "Malloc disk", 00:20:51.968 "block_size": 512, 00:20:51.968 "num_blocks": 65536, 00:20:51.968 "uuid": "6d936500-88a8-4bb2-9901-c2f5866c6f8d", 00:20:51.968 "assigned_rate_limits": { 00:20:51.968 "rw_ios_per_sec": 0, 00:20:51.968 "rw_mbytes_per_sec": 0, 00:20:51.968 "r_mbytes_per_sec": 0, 00:20:51.968 "w_mbytes_per_sec": 0 00:20:51.968 }, 00:20:51.968 "claimed": true, 00:20:51.968 "claim_type": "exclusive_write", 00:20:51.968 "zoned": false, 00:20:51.968 "supported_io_types": { 00:20:51.968 "read": true, 00:20:51.968 "write": true, 00:20:51.968 "unmap": true, 00:20:51.968 "flush": true, 00:20:51.968 "reset": true, 00:20:51.968 "nvme_admin": false, 00:20:51.968 "nvme_io": false, 00:20:51.968 "nvme_io_md": false, 00:20:51.968 "write_zeroes": true, 00:20:51.968 "zcopy": true, 00:20:51.968 "get_zone_info": false, 00:20:51.968 "zone_management": false, 00:20:51.968 "zone_append": false, 00:20:51.968 "compare": false, 00:20:51.968 "compare_and_write": false, 00:20:51.968 "abort": true, 00:20:51.968 "seek_hole": false, 00:20:51.968 "seek_data": false, 00:20:51.968 "copy": true, 00:20:51.968 "nvme_iov_md": false 00:20:51.968 }, 00:20:51.968 "memory_domains": [ 00:20:51.968 { 00:20:51.968 "dma_device_id": "system", 00:20:51.968 "dma_device_type": 1 00:20:51.968 }, 00:20:51.968 { 00:20:51.968 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:51.968 "dma_device_type": 2 00:20:51.968 } 00:20:51.968 ], 00:20:51.968 "driver_specific": {} 00:20:51.968 } 00:20:51.968 ] 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:51.968 20:19:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:51.968 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:51.968 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:51.968 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:51.969 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:51.969 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:51.969 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:51.969 "name": "Existed_Raid", 00:20:51.969 "uuid": "f683ffd2-66da-437f-ab61-50e4c3aeee31", 00:20:51.969 "strip_size_kb": 0, 00:20:51.969 "state": "online", 00:20:51.969 "raid_level": "raid1", 00:20:51.969 "superblock": false, 00:20:51.969 "num_base_bdevs": 2, 00:20:51.969 "num_base_bdevs_discovered": 2, 00:20:51.969 "num_base_bdevs_operational": 2, 00:20:51.969 "base_bdevs_list": [ 00:20:51.969 { 00:20:51.969 "name": "BaseBdev1", 00:20:51.969 "uuid": "5b53a3c9-5e50-474d-a76c-293c38332991", 00:20:51.969 "is_configured": true, 00:20:51.969 "data_offset": 0, 00:20:51.969 "data_size": 65536 00:20:51.969 }, 00:20:51.969 { 00:20:51.969 "name": "BaseBdev2", 00:20:51.969 "uuid": "6d936500-88a8-4bb2-9901-c2f5866c6f8d", 00:20:51.969 "is_configured": true, 00:20:51.969 "data_offset": 0, 00:20:51.969 "data_size": 65536 00:20:51.969 } 00:20:51.969 ] 00:20:51.969 }' 00:20:51.969 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:51.969 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:52.534 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:20:52.534 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:20:52.534 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:20:52.534 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:20:52.534 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:20:52.534 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:20:52.534 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:20:52.534 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:20:52.534 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:52.534 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:52.534 [2024-10-01 20:19:47.536483] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:52.534 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:52.534 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:20:52.534 "name": "Existed_Raid", 00:20:52.534 "aliases": [ 00:20:52.534 "f683ffd2-66da-437f-ab61-50e4c3aeee31" 00:20:52.534 ], 00:20:52.534 "product_name": "Raid Volume", 00:20:52.534 "block_size": 512, 00:20:52.534 "num_blocks": 65536, 00:20:52.534 "uuid": "f683ffd2-66da-437f-ab61-50e4c3aeee31", 00:20:52.534 "assigned_rate_limits": { 00:20:52.534 "rw_ios_per_sec": 0, 00:20:52.534 "rw_mbytes_per_sec": 0, 00:20:52.534 "r_mbytes_per_sec": 0, 00:20:52.534 "w_mbytes_per_sec": 0 00:20:52.534 }, 00:20:52.534 "claimed": false, 00:20:52.534 "zoned": false, 00:20:52.534 "supported_io_types": { 00:20:52.534 "read": true, 00:20:52.534 "write": true, 00:20:52.534 "unmap": false, 00:20:52.534 "flush": false, 00:20:52.534 "reset": true, 00:20:52.534 "nvme_admin": false, 00:20:52.534 "nvme_io": false, 00:20:52.534 "nvme_io_md": false, 00:20:52.534 "write_zeroes": true, 00:20:52.534 "zcopy": false, 00:20:52.534 "get_zone_info": false, 00:20:52.534 "zone_management": false, 00:20:52.534 "zone_append": false, 00:20:52.534 "compare": false, 00:20:52.534 "compare_and_write": false, 00:20:52.534 "abort": false, 00:20:52.534 "seek_hole": false, 00:20:52.534 "seek_data": false, 00:20:52.534 "copy": false, 00:20:52.534 "nvme_iov_md": false 00:20:52.534 }, 00:20:52.534 "memory_domains": [ 00:20:52.534 { 00:20:52.534 "dma_device_id": "system", 00:20:52.534 "dma_device_type": 1 00:20:52.534 }, 00:20:52.534 { 00:20:52.534 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:52.534 "dma_device_type": 2 00:20:52.534 }, 00:20:52.534 { 00:20:52.534 "dma_device_id": "system", 00:20:52.534 "dma_device_type": 1 00:20:52.534 }, 00:20:52.534 { 00:20:52.534 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:52.534 "dma_device_type": 2 00:20:52.534 } 00:20:52.534 ], 00:20:52.534 "driver_specific": { 00:20:52.534 "raid": { 00:20:52.534 "uuid": "f683ffd2-66da-437f-ab61-50e4c3aeee31", 00:20:52.534 "strip_size_kb": 0, 00:20:52.534 "state": "online", 00:20:52.534 "raid_level": "raid1", 00:20:52.535 "superblock": false, 00:20:52.535 "num_base_bdevs": 2, 00:20:52.535 "num_base_bdevs_discovered": 2, 00:20:52.535 "num_base_bdevs_operational": 2, 00:20:52.535 "base_bdevs_list": [ 00:20:52.535 { 00:20:52.535 "name": "BaseBdev1", 00:20:52.535 "uuid": "5b53a3c9-5e50-474d-a76c-293c38332991", 00:20:52.535 "is_configured": true, 00:20:52.535 "data_offset": 0, 00:20:52.535 "data_size": 65536 00:20:52.535 }, 00:20:52.535 { 00:20:52.535 "name": "BaseBdev2", 00:20:52.535 "uuid": "6d936500-88a8-4bb2-9901-c2f5866c6f8d", 00:20:52.535 "is_configured": true, 00:20:52.535 "data_offset": 0, 00:20:52.535 "data_size": 65536 00:20:52.535 } 00:20:52.535 ] 00:20:52.535 } 00:20:52.535 } 00:20:52.535 }' 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:20:52.535 BaseBdev2' 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:52.535 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:52.793 [2024-10-01 20:19:47.832369] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:52.793 "name": "Existed_Raid", 00:20:52.793 "uuid": "f683ffd2-66da-437f-ab61-50e4c3aeee31", 00:20:52.793 "strip_size_kb": 0, 00:20:52.793 "state": "online", 00:20:52.793 "raid_level": "raid1", 00:20:52.793 "superblock": false, 00:20:52.793 "num_base_bdevs": 2, 00:20:52.793 "num_base_bdevs_discovered": 1, 00:20:52.793 "num_base_bdevs_operational": 1, 00:20:52.793 "base_bdevs_list": [ 00:20:52.793 { 00:20:52.793 "name": null, 00:20:52.793 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:52.793 "is_configured": false, 00:20:52.793 "data_offset": 0, 00:20:52.793 "data_size": 65536 00:20:52.793 }, 00:20:52.793 { 00:20:52.793 "name": "BaseBdev2", 00:20:52.793 "uuid": "6d936500-88a8-4bb2-9901-c2f5866c6f8d", 00:20:52.793 "is_configured": true, 00:20:52.793 "data_offset": 0, 00:20:52.793 "data_size": 65536 00:20:52.793 } 00:20:52.793 ] 00:20:52.793 }' 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:52.793 20:19:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:53.361 [2024-10-01 20:19:48.497230] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:20:53.361 [2024-10-01 20:19:48.497555] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:20:53.361 [2024-10-01 20:19:48.576748] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:53.361 [2024-10-01 20:19:48.576871] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:20:53.361 [2024-10-01 20:19:48.576893] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:53.361 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:53.620 20:19:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:20:53.620 20:19:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:20:53.620 20:19:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:20:53.620 20:19:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 63126 00:20:53.620 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 63126 ']' 00:20:53.620 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 63126 00:20:53.620 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:20:53.620 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:20:53.620 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 63126 00:20:53.620 killing process with pid 63126 00:20:53.620 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:20:53.620 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:20:53.620 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 63126' 00:20:53.620 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 63126 00:20:53.620 [2024-10-01 20:19:48.671069] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:20:53.620 20:19:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 63126 00:20:53.620 [2024-10-01 20:19:48.685989] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:20:55.524 00:20:55.524 real 0m6.396s 00:20:55.524 user 0m9.164s 00:20:55.524 sys 0m0.918s 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:20:55.524 ************************************ 00:20:55.524 END TEST raid_state_function_test 00:20:55.524 ************************************ 00:20:55.524 20:19:50 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 2 true 00:20:55.524 20:19:50 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:20:55.524 20:19:50 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:20:55.524 20:19:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:20:55.524 ************************************ 00:20:55.524 START TEST raid_state_function_test_sb 00:20:55.524 ************************************ 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 true 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:20:55.524 Process raid pid: 63396 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=63396 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 63396' 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 63396 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 63396 ']' 00:20:55.524 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:55.524 20:19:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:55.524 [2024-10-01 20:19:50.622460] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:20:55.524 [2024-10-01 20:19:50.622969] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:55.782 [2024-10-01 20:19:50.803878] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:56.041 [2024-10-01 20:19:51.041139] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:20:56.041 [2024-10-01 20:19:51.243908] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:56.041 [2024-10-01 20:19:51.243973] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:56.609 [2024-10-01 20:19:51.674528] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:20:56.609 [2024-10-01 20:19:51.674611] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:20:56.609 [2024-10-01 20:19:51.674628] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:20:56.609 [2024-10-01 20:19:51.674645] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:56.609 "name": "Existed_Raid", 00:20:56.609 "uuid": "327a8f14-9f1e-4cf0-a693-7535f62d8de9", 00:20:56.609 "strip_size_kb": 0, 00:20:56.609 "state": "configuring", 00:20:56.609 "raid_level": "raid1", 00:20:56.609 "superblock": true, 00:20:56.609 "num_base_bdevs": 2, 00:20:56.609 "num_base_bdevs_discovered": 0, 00:20:56.609 "num_base_bdevs_operational": 2, 00:20:56.609 "base_bdevs_list": [ 00:20:56.609 { 00:20:56.609 "name": "BaseBdev1", 00:20:56.609 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:56.609 "is_configured": false, 00:20:56.609 "data_offset": 0, 00:20:56.609 "data_size": 0 00:20:56.609 }, 00:20:56.609 { 00:20:56.609 "name": "BaseBdev2", 00:20:56.609 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:56.609 "is_configured": false, 00:20:56.609 "data_offset": 0, 00:20:56.609 "data_size": 0 00:20:56.609 } 00:20:56.609 ] 00:20:56.609 }' 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:56.609 20:19:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:57.176 [2024-10-01 20:19:52.198573] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:20:57.176 [2024-10-01 20:19:52.198619] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:57.176 [2024-10-01 20:19:52.206581] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:20:57.176 [2024-10-01 20:19:52.206651] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:20:57.176 [2024-10-01 20:19:52.206666] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:20:57.176 [2024-10-01 20:19:52.206685] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:57.176 [2024-10-01 20:19:52.251877] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:20:57.176 BaseBdev1 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:57.176 [ 00:20:57.176 { 00:20:57.176 "name": "BaseBdev1", 00:20:57.176 "aliases": [ 00:20:57.176 "a3682f10-b3d8-498a-81df-2d1b0ecd40da" 00:20:57.176 ], 00:20:57.176 "product_name": "Malloc disk", 00:20:57.176 "block_size": 512, 00:20:57.176 "num_blocks": 65536, 00:20:57.176 "uuid": "a3682f10-b3d8-498a-81df-2d1b0ecd40da", 00:20:57.176 "assigned_rate_limits": { 00:20:57.176 "rw_ios_per_sec": 0, 00:20:57.176 "rw_mbytes_per_sec": 0, 00:20:57.176 "r_mbytes_per_sec": 0, 00:20:57.176 "w_mbytes_per_sec": 0 00:20:57.176 }, 00:20:57.176 "claimed": true, 00:20:57.176 "claim_type": "exclusive_write", 00:20:57.176 "zoned": false, 00:20:57.176 "supported_io_types": { 00:20:57.176 "read": true, 00:20:57.176 "write": true, 00:20:57.176 "unmap": true, 00:20:57.176 "flush": true, 00:20:57.176 "reset": true, 00:20:57.176 "nvme_admin": false, 00:20:57.176 "nvme_io": false, 00:20:57.176 "nvme_io_md": false, 00:20:57.176 "write_zeroes": true, 00:20:57.176 "zcopy": true, 00:20:57.176 "get_zone_info": false, 00:20:57.176 "zone_management": false, 00:20:57.176 "zone_append": false, 00:20:57.176 "compare": false, 00:20:57.176 "compare_and_write": false, 00:20:57.176 "abort": true, 00:20:57.176 "seek_hole": false, 00:20:57.176 "seek_data": false, 00:20:57.176 "copy": true, 00:20:57.176 "nvme_iov_md": false 00:20:57.176 }, 00:20:57.176 "memory_domains": [ 00:20:57.176 { 00:20:57.176 "dma_device_id": "system", 00:20:57.176 "dma_device_type": 1 00:20:57.176 }, 00:20:57.176 { 00:20:57.176 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:57.176 "dma_device_type": 2 00:20:57.176 } 00:20:57.176 ], 00:20:57.176 "driver_specific": {} 00:20:57.176 } 00:20:57.176 ] 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:57.176 "name": "Existed_Raid", 00:20:57.176 "uuid": "2715cb68-a172-4d0f-8f31-416ffe23d2fd", 00:20:57.176 "strip_size_kb": 0, 00:20:57.176 "state": "configuring", 00:20:57.176 "raid_level": "raid1", 00:20:57.176 "superblock": true, 00:20:57.176 "num_base_bdevs": 2, 00:20:57.176 "num_base_bdevs_discovered": 1, 00:20:57.176 "num_base_bdevs_operational": 2, 00:20:57.176 "base_bdevs_list": [ 00:20:57.176 { 00:20:57.176 "name": "BaseBdev1", 00:20:57.176 "uuid": "a3682f10-b3d8-498a-81df-2d1b0ecd40da", 00:20:57.176 "is_configured": true, 00:20:57.176 "data_offset": 2048, 00:20:57.176 "data_size": 63488 00:20:57.176 }, 00:20:57.176 { 00:20:57.176 "name": "BaseBdev2", 00:20:57.176 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:57.176 "is_configured": false, 00:20:57.176 "data_offset": 0, 00:20:57.176 "data_size": 0 00:20:57.176 } 00:20:57.176 ] 00:20:57.176 }' 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:57.176 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:57.744 [2024-10-01 20:19:52.780174] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:20:57.744 [2024-10-01 20:19:52.780237] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:57.744 [2024-10-01 20:19:52.788201] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:20:57.744 [2024-10-01 20:19:52.790903] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:20:57.744 [2024-10-01 20:19:52.790972] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:57.744 "name": "Existed_Raid", 00:20:57.744 "uuid": "18bd5bd4-9424-44b2-a9fc-ea273eba980d", 00:20:57.744 "strip_size_kb": 0, 00:20:57.744 "state": "configuring", 00:20:57.744 "raid_level": "raid1", 00:20:57.744 "superblock": true, 00:20:57.744 "num_base_bdevs": 2, 00:20:57.744 "num_base_bdevs_discovered": 1, 00:20:57.744 "num_base_bdevs_operational": 2, 00:20:57.744 "base_bdevs_list": [ 00:20:57.744 { 00:20:57.744 "name": "BaseBdev1", 00:20:57.744 "uuid": "a3682f10-b3d8-498a-81df-2d1b0ecd40da", 00:20:57.744 "is_configured": true, 00:20:57.744 "data_offset": 2048, 00:20:57.744 "data_size": 63488 00:20:57.744 }, 00:20:57.744 { 00:20:57.744 "name": "BaseBdev2", 00:20:57.744 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:57.744 "is_configured": false, 00:20:57.744 "data_offset": 0, 00:20:57.744 "data_size": 0 00:20:57.744 } 00:20:57.744 ] 00:20:57.744 }' 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:57.744 20:19:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:58.322 [2024-10-01 20:19:53.347581] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:20:58.322 [2024-10-01 20:19:53.347998] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:20:58.322 [2024-10-01 20:19:53.348021] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:20:58.322 [2024-10-01 20:19:53.348379] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:20:58.322 BaseBdev2 00:20:58.322 [2024-10-01 20:19:53.348570] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:20:58.322 [2024-10-01 20:19:53.348591] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:20:58.322 [2024-10-01 20:19:53.348794] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:58.322 [ 00:20:58.322 { 00:20:58.322 "name": "BaseBdev2", 00:20:58.322 "aliases": [ 00:20:58.322 "9f872967-72ed-469b-bdd4-fc456786fa67" 00:20:58.322 ], 00:20:58.322 "product_name": "Malloc disk", 00:20:58.322 "block_size": 512, 00:20:58.322 "num_blocks": 65536, 00:20:58.322 "uuid": "9f872967-72ed-469b-bdd4-fc456786fa67", 00:20:58.322 "assigned_rate_limits": { 00:20:58.322 "rw_ios_per_sec": 0, 00:20:58.322 "rw_mbytes_per_sec": 0, 00:20:58.322 "r_mbytes_per_sec": 0, 00:20:58.322 "w_mbytes_per_sec": 0 00:20:58.322 }, 00:20:58.322 "claimed": true, 00:20:58.322 "claim_type": "exclusive_write", 00:20:58.322 "zoned": false, 00:20:58.322 "supported_io_types": { 00:20:58.322 "read": true, 00:20:58.322 "write": true, 00:20:58.322 "unmap": true, 00:20:58.322 "flush": true, 00:20:58.322 "reset": true, 00:20:58.322 "nvme_admin": false, 00:20:58.322 "nvme_io": false, 00:20:58.322 "nvme_io_md": false, 00:20:58.322 "write_zeroes": true, 00:20:58.322 "zcopy": true, 00:20:58.322 "get_zone_info": false, 00:20:58.322 "zone_management": false, 00:20:58.322 "zone_append": false, 00:20:58.322 "compare": false, 00:20:58.322 "compare_and_write": false, 00:20:58.322 "abort": true, 00:20:58.322 "seek_hole": false, 00:20:58.322 "seek_data": false, 00:20:58.322 "copy": true, 00:20:58.322 "nvme_iov_md": false 00:20:58.322 }, 00:20:58.322 "memory_domains": [ 00:20:58.322 { 00:20:58.322 "dma_device_id": "system", 00:20:58.322 "dma_device_type": 1 00:20:58.322 }, 00:20:58.322 { 00:20:58.322 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:58.322 "dma_device_type": 2 00:20:58.322 } 00:20:58.322 ], 00:20:58.322 "driver_specific": {} 00:20:58.322 } 00:20:58.322 ] 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:58.322 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:58.322 "name": "Existed_Raid", 00:20:58.322 "uuid": "18bd5bd4-9424-44b2-a9fc-ea273eba980d", 00:20:58.322 "strip_size_kb": 0, 00:20:58.323 "state": "online", 00:20:58.323 "raid_level": "raid1", 00:20:58.323 "superblock": true, 00:20:58.323 "num_base_bdevs": 2, 00:20:58.323 "num_base_bdevs_discovered": 2, 00:20:58.323 "num_base_bdevs_operational": 2, 00:20:58.323 "base_bdevs_list": [ 00:20:58.323 { 00:20:58.323 "name": "BaseBdev1", 00:20:58.323 "uuid": "a3682f10-b3d8-498a-81df-2d1b0ecd40da", 00:20:58.323 "is_configured": true, 00:20:58.323 "data_offset": 2048, 00:20:58.323 "data_size": 63488 00:20:58.323 }, 00:20:58.323 { 00:20:58.323 "name": "BaseBdev2", 00:20:58.323 "uuid": "9f872967-72ed-469b-bdd4-fc456786fa67", 00:20:58.323 "is_configured": true, 00:20:58.323 "data_offset": 2048, 00:20:58.323 "data_size": 63488 00:20:58.323 } 00:20:58.323 ] 00:20:58.323 }' 00:20:58.323 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:58.323 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:58.898 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:20:58.898 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:20:58.898 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:20:58.898 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:20:58.898 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:20:58.898 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:20:58.898 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:20:58.898 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:20:58.898 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:58.898 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:58.898 [2024-10-01 20:19:53.908344] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:20:58.898 20:19:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:58.898 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:20:58.898 "name": "Existed_Raid", 00:20:58.898 "aliases": [ 00:20:58.898 "18bd5bd4-9424-44b2-a9fc-ea273eba980d" 00:20:58.898 ], 00:20:58.898 "product_name": "Raid Volume", 00:20:58.898 "block_size": 512, 00:20:58.898 "num_blocks": 63488, 00:20:58.898 "uuid": "18bd5bd4-9424-44b2-a9fc-ea273eba980d", 00:20:58.898 "assigned_rate_limits": { 00:20:58.898 "rw_ios_per_sec": 0, 00:20:58.898 "rw_mbytes_per_sec": 0, 00:20:58.898 "r_mbytes_per_sec": 0, 00:20:58.898 "w_mbytes_per_sec": 0 00:20:58.898 }, 00:20:58.898 "claimed": false, 00:20:58.898 "zoned": false, 00:20:58.898 "supported_io_types": { 00:20:58.898 "read": true, 00:20:58.898 "write": true, 00:20:58.898 "unmap": false, 00:20:58.898 "flush": false, 00:20:58.898 "reset": true, 00:20:58.898 "nvme_admin": false, 00:20:58.898 "nvme_io": false, 00:20:58.898 "nvme_io_md": false, 00:20:58.898 "write_zeroes": true, 00:20:58.898 "zcopy": false, 00:20:58.898 "get_zone_info": false, 00:20:58.898 "zone_management": false, 00:20:58.898 "zone_append": false, 00:20:58.898 "compare": false, 00:20:58.898 "compare_and_write": false, 00:20:58.898 "abort": false, 00:20:58.898 "seek_hole": false, 00:20:58.898 "seek_data": false, 00:20:58.898 "copy": false, 00:20:58.898 "nvme_iov_md": false 00:20:58.898 }, 00:20:58.898 "memory_domains": [ 00:20:58.898 { 00:20:58.898 "dma_device_id": "system", 00:20:58.898 "dma_device_type": 1 00:20:58.898 }, 00:20:58.898 { 00:20:58.898 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:58.898 "dma_device_type": 2 00:20:58.898 }, 00:20:58.898 { 00:20:58.898 "dma_device_id": "system", 00:20:58.898 "dma_device_type": 1 00:20:58.898 }, 00:20:58.898 { 00:20:58.898 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:20:58.898 "dma_device_type": 2 00:20:58.898 } 00:20:58.898 ], 00:20:58.898 "driver_specific": { 00:20:58.898 "raid": { 00:20:58.898 "uuid": "18bd5bd4-9424-44b2-a9fc-ea273eba980d", 00:20:58.898 "strip_size_kb": 0, 00:20:58.898 "state": "online", 00:20:58.898 "raid_level": "raid1", 00:20:58.898 "superblock": true, 00:20:58.898 "num_base_bdevs": 2, 00:20:58.898 "num_base_bdevs_discovered": 2, 00:20:58.898 "num_base_bdevs_operational": 2, 00:20:58.898 "base_bdevs_list": [ 00:20:58.898 { 00:20:58.898 "name": "BaseBdev1", 00:20:58.898 "uuid": "a3682f10-b3d8-498a-81df-2d1b0ecd40da", 00:20:58.898 "is_configured": true, 00:20:58.898 "data_offset": 2048, 00:20:58.898 "data_size": 63488 00:20:58.898 }, 00:20:58.898 { 00:20:58.898 "name": "BaseBdev2", 00:20:58.898 "uuid": "9f872967-72ed-469b-bdd4-fc456786fa67", 00:20:58.898 "is_configured": true, 00:20:58.898 "data_offset": 2048, 00:20:58.898 "data_size": 63488 00:20:58.898 } 00:20:58.898 ] 00:20:58.898 } 00:20:58.898 } 00:20:58.898 }' 00:20:58.898 20:19:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:20:58.898 BaseBdev2' 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:58.898 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:59.157 [2024-10-01 20:19:54.180051] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:20:59.157 "name": "Existed_Raid", 00:20:59.157 "uuid": "18bd5bd4-9424-44b2-a9fc-ea273eba980d", 00:20:59.157 "strip_size_kb": 0, 00:20:59.157 "state": "online", 00:20:59.157 "raid_level": "raid1", 00:20:59.157 "superblock": true, 00:20:59.157 "num_base_bdevs": 2, 00:20:59.157 "num_base_bdevs_discovered": 1, 00:20:59.157 "num_base_bdevs_operational": 1, 00:20:59.157 "base_bdevs_list": [ 00:20:59.157 { 00:20:59.157 "name": null, 00:20:59.157 "uuid": "00000000-0000-0000-0000-000000000000", 00:20:59.157 "is_configured": false, 00:20:59.157 "data_offset": 0, 00:20:59.157 "data_size": 63488 00:20:59.157 }, 00:20:59.157 { 00:20:59.157 "name": "BaseBdev2", 00:20:59.157 "uuid": "9f872967-72ed-469b-bdd4-fc456786fa67", 00:20:59.157 "is_configured": true, 00:20:59.157 "data_offset": 2048, 00:20:59.157 "data_size": 63488 00:20:59.157 } 00:20:59.157 ] 00:20:59.157 }' 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:20:59.157 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:59.722 [2024-10-01 20:19:54.850247] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:20:59.722 [2024-10-01 20:19:54.850525] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:20:59.722 [2024-10-01 20:19:54.937500] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:20:59.722 [2024-10-01 20:19:54.937582] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:20:59.722 [2024-10-01 20:19:54.937603] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:20:59.722 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:59.980 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:20:59.980 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:20:59.980 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:20:59.980 20:19:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 63396 00:20:59.980 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 63396 ']' 00:20:59.980 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 63396 00:20:59.980 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:20:59.980 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:20:59.980 20:19:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 63396 00:20:59.980 killing process with pid 63396 00:20:59.980 20:19:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:20:59.980 20:19:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:20:59.980 20:19:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 63396' 00:20:59.980 20:19:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 63396 00:20:59.980 [2024-10-01 20:19:55.030513] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:20:59.980 20:19:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 63396 00:20:59.980 [2024-10-01 20:19:55.045950] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:21:01.927 ************************************ 00:21:01.927 END TEST raid_state_function_test_sb 00:21:01.927 ************************************ 00:21:01.927 20:19:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:21:01.927 00:21:01.927 real 0m6.368s 00:21:01.927 user 0m9.058s 00:21:01.927 sys 0m0.947s 00:21:01.927 20:19:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:21:01.927 20:19:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:01.927 20:19:56 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 2 00:21:01.927 20:19:56 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:21:01.927 20:19:56 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:21:01.927 20:19:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:21:01.927 ************************************ 00:21:01.927 START TEST raid_superblock_test 00:21:01.927 ************************************ 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 2 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=63659 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 63659 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 63659 ']' 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:21:01.927 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:21:01.927 20:19:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:01.927 [2024-10-01 20:19:57.036345] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:21:01.927 [2024-10-01 20:19:57.036812] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid63659 ] 00:21:02.185 [2024-10-01 20:19:57.217601] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:02.443 [2024-10-01 20:19:57.471594] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:21:02.443 [2024-10-01 20:19:57.674661] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:02.443 [2024-10-01 20:19:57.674711] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:03.009 malloc1 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:03.009 [2024-10-01 20:19:58.126192] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:21:03.009 [2024-10-01 20:19:58.126476] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:03.009 [2024-10-01 20:19:58.126523] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:21:03.009 [2024-10-01 20:19:58.126545] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:03.009 [2024-10-01 20:19:58.129496] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:03.009 [2024-10-01 20:19:58.129699] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:21:03.009 pt1 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:03.009 malloc2 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:03.009 [2024-10-01 20:19:58.181869] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:21:03.009 [2024-10-01 20:19:58.182106] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:03.009 [2024-10-01 20:19:58.182157] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:21:03.009 [2024-10-01 20:19:58.182174] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:03.009 [2024-10-01 20:19:58.185085] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:03.009 [2024-10-01 20:19:58.185260] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:21:03.009 pt2 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:03.009 [2024-10-01 20:19:58.193974] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:21:03.009 [2024-10-01 20:19:58.196677] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:21:03.009 [2024-10-01 20:19:58.197010] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:21:03.009 [2024-10-01 20:19:58.197031] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:21:03.009 [2024-10-01 20:19:58.197356] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:21:03.009 [2024-10-01 20:19:58.197593] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:21:03.009 [2024-10-01 20:19:58.197614] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:21:03.009 [2024-10-01 20:19:58.197939] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:03.009 "name": "raid_bdev1", 00:21:03.009 "uuid": "8b33d87d-9289-438d-9414-70be971e2f32", 00:21:03.009 "strip_size_kb": 0, 00:21:03.009 "state": "online", 00:21:03.009 "raid_level": "raid1", 00:21:03.009 "superblock": true, 00:21:03.009 "num_base_bdevs": 2, 00:21:03.009 "num_base_bdevs_discovered": 2, 00:21:03.009 "num_base_bdevs_operational": 2, 00:21:03.009 "base_bdevs_list": [ 00:21:03.009 { 00:21:03.009 "name": "pt1", 00:21:03.009 "uuid": "00000000-0000-0000-0000-000000000001", 00:21:03.009 "is_configured": true, 00:21:03.009 "data_offset": 2048, 00:21:03.009 "data_size": 63488 00:21:03.009 }, 00:21:03.009 { 00:21:03.009 "name": "pt2", 00:21:03.009 "uuid": "00000000-0000-0000-0000-000000000002", 00:21:03.009 "is_configured": true, 00:21:03.009 "data_offset": 2048, 00:21:03.009 "data_size": 63488 00:21:03.009 } 00:21:03.009 ] 00:21:03.009 }' 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:03.009 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:03.575 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:21:03.575 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:21:03.575 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:21:03.575 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:21:03.575 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:21:03.575 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:21:03.575 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:21:03.575 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:03.575 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:21:03.575 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:03.575 [2024-10-01 20:19:58.702523] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:21:03.575 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:03.575 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:21:03.575 "name": "raid_bdev1", 00:21:03.575 "aliases": [ 00:21:03.575 "8b33d87d-9289-438d-9414-70be971e2f32" 00:21:03.575 ], 00:21:03.575 "product_name": "Raid Volume", 00:21:03.575 "block_size": 512, 00:21:03.575 "num_blocks": 63488, 00:21:03.575 "uuid": "8b33d87d-9289-438d-9414-70be971e2f32", 00:21:03.575 "assigned_rate_limits": { 00:21:03.575 "rw_ios_per_sec": 0, 00:21:03.575 "rw_mbytes_per_sec": 0, 00:21:03.575 "r_mbytes_per_sec": 0, 00:21:03.575 "w_mbytes_per_sec": 0 00:21:03.575 }, 00:21:03.575 "claimed": false, 00:21:03.575 "zoned": false, 00:21:03.575 "supported_io_types": { 00:21:03.575 "read": true, 00:21:03.575 "write": true, 00:21:03.575 "unmap": false, 00:21:03.575 "flush": false, 00:21:03.575 "reset": true, 00:21:03.575 "nvme_admin": false, 00:21:03.575 "nvme_io": false, 00:21:03.575 "nvme_io_md": false, 00:21:03.575 "write_zeroes": true, 00:21:03.575 "zcopy": false, 00:21:03.575 "get_zone_info": false, 00:21:03.575 "zone_management": false, 00:21:03.575 "zone_append": false, 00:21:03.575 "compare": false, 00:21:03.575 "compare_and_write": false, 00:21:03.575 "abort": false, 00:21:03.575 "seek_hole": false, 00:21:03.575 "seek_data": false, 00:21:03.575 "copy": false, 00:21:03.575 "nvme_iov_md": false 00:21:03.575 }, 00:21:03.575 "memory_domains": [ 00:21:03.575 { 00:21:03.575 "dma_device_id": "system", 00:21:03.575 "dma_device_type": 1 00:21:03.575 }, 00:21:03.575 { 00:21:03.575 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:03.575 "dma_device_type": 2 00:21:03.575 }, 00:21:03.575 { 00:21:03.575 "dma_device_id": "system", 00:21:03.575 "dma_device_type": 1 00:21:03.575 }, 00:21:03.575 { 00:21:03.575 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:03.575 "dma_device_type": 2 00:21:03.575 } 00:21:03.575 ], 00:21:03.575 "driver_specific": { 00:21:03.575 "raid": { 00:21:03.575 "uuid": "8b33d87d-9289-438d-9414-70be971e2f32", 00:21:03.575 "strip_size_kb": 0, 00:21:03.575 "state": "online", 00:21:03.575 "raid_level": "raid1", 00:21:03.575 "superblock": true, 00:21:03.575 "num_base_bdevs": 2, 00:21:03.575 "num_base_bdevs_discovered": 2, 00:21:03.575 "num_base_bdevs_operational": 2, 00:21:03.575 "base_bdevs_list": [ 00:21:03.575 { 00:21:03.575 "name": "pt1", 00:21:03.575 "uuid": "00000000-0000-0000-0000-000000000001", 00:21:03.575 "is_configured": true, 00:21:03.575 "data_offset": 2048, 00:21:03.575 "data_size": 63488 00:21:03.575 }, 00:21:03.575 { 00:21:03.575 "name": "pt2", 00:21:03.575 "uuid": "00000000-0000-0000-0000-000000000002", 00:21:03.575 "is_configured": true, 00:21:03.575 "data_offset": 2048, 00:21:03.575 "data_size": 63488 00:21:03.575 } 00:21:03.575 ] 00:21:03.575 } 00:21:03.575 } 00:21:03.575 }' 00:21:03.575 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:21:03.576 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:21:03.576 pt2' 00:21:03.576 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:03.833 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:21:03.833 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:03.833 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:21:03.833 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:03.833 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:03.833 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:03.833 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:03.833 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:03.833 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:03.833 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:03.833 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:21:03.833 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:03.833 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:03.833 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:03.834 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:03.834 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:03.834 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:03.834 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:21:03.834 20:19:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:21:03.834 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:03.834 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:03.834 [2024-10-01 20:19:58.970572] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:21:03.834 20:19:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=8b33d87d-9289-438d-9414-70be971e2f32 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 8b33d87d-9289-438d-9414-70be971e2f32 ']' 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:03.834 [2024-10-01 20:19:59.018270] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:21:03.834 [2024-10-01 20:19:59.018303] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:21:03.834 [2024-10-01 20:19:59.018406] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:21:03.834 [2024-10-01 20:19:59.018506] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:21:03.834 [2024-10-01 20:19:59.018526] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:03.834 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:04.092 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:04.092 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:21:04.092 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:21:04.092 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:04.092 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:04.092 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:04.092 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:21:04.092 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:04.093 [2024-10-01 20:19:59.158263] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:21:04.093 [2024-10-01 20:19:59.161115] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:21:04.093 [2024-10-01 20:19:59.161253] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:21:04.093 [2024-10-01 20:19:59.161356] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:21:04.093 [2024-10-01 20:19:59.161381] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:21:04.093 [2024-10-01 20:19:59.161396] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:21:04.093 request: 00:21:04.093 { 00:21:04.093 "name": "raid_bdev1", 00:21:04.093 "raid_level": "raid1", 00:21:04.093 "base_bdevs": [ 00:21:04.093 "malloc1", 00:21:04.093 "malloc2" 00:21:04.093 ], 00:21:04.093 "superblock": false, 00:21:04.093 "method": "bdev_raid_create", 00:21:04.093 "req_id": 1 00:21:04.093 } 00:21:04.093 Got JSON-RPC error response 00:21:04.093 response: 00:21:04.093 { 00:21:04.093 "code": -17, 00:21:04.093 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:21:04.093 } 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:04.093 [2024-10-01 20:19:59.226343] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:21:04.093 [2024-10-01 20:19:59.226600] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:04.093 [2024-10-01 20:19:59.226644] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:21:04.093 [2024-10-01 20:19:59.226665] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:04.093 [2024-10-01 20:19:59.229763] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:04.093 [2024-10-01 20:19:59.229843] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:21:04.093 [2024-10-01 20:19:59.229952] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:21:04.093 [2024-10-01 20:19:59.230029] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:21:04.093 pt1 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:04.093 "name": "raid_bdev1", 00:21:04.093 "uuid": "8b33d87d-9289-438d-9414-70be971e2f32", 00:21:04.093 "strip_size_kb": 0, 00:21:04.093 "state": "configuring", 00:21:04.093 "raid_level": "raid1", 00:21:04.093 "superblock": true, 00:21:04.093 "num_base_bdevs": 2, 00:21:04.093 "num_base_bdevs_discovered": 1, 00:21:04.093 "num_base_bdevs_operational": 2, 00:21:04.093 "base_bdevs_list": [ 00:21:04.093 { 00:21:04.093 "name": "pt1", 00:21:04.093 "uuid": "00000000-0000-0000-0000-000000000001", 00:21:04.093 "is_configured": true, 00:21:04.093 "data_offset": 2048, 00:21:04.093 "data_size": 63488 00:21:04.093 }, 00:21:04.093 { 00:21:04.093 "name": null, 00:21:04.093 "uuid": "00000000-0000-0000-0000-000000000002", 00:21:04.093 "is_configured": false, 00:21:04.093 "data_offset": 2048, 00:21:04.093 "data_size": 63488 00:21:04.093 } 00:21:04.093 ] 00:21:04.093 }' 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:04.093 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:04.658 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:21:04.658 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:21:04.658 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:21:04.658 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:04.659 [2024-10-01 20:19:59.746640] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:21:04.659 [2024-10-01 20:19:59.746896] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:04.659 [2024-10-01 20:19:59.746951] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:21:04.659 [2024-10-01 20:19:59.746972] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:04.659 [2024-10-01 20:19:59.747653] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:04.659 [2024-10-01 20:19:59.747691] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:21:04.659 [2024-10-01 20:19:59.747848] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:21:04.659 [2024-10-01 20:19:59.747887] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:21:04.659 [2024-10-01 20:19:59.748038] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:21:04.659 [2024-10-01 20:19:59.748060] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:21:04.659 [2024-10-01 20:19:59.748375] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:21:04.659 [2024-10-01 20:19:59.748607] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:21:04.659 [2024-10-01 20:19:59.748623] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:21:04.659 [2024-10-01 20:19:59.748827] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:04.659 pt2 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:04.659 "name": "raid_bdev1", 00:21:04.659 "uuid": "8b33d87d-9289-438d-9414-70be971e2f32", 00:21:04.659 "strip_size_kb": 0, 00:21:04.659 "state": "online", 00:21:04.659 "raid_level": "raid1", 00:21:04.659 "superblock": true, 00:21:04.659 "num_base_bdevs": 2, 00:21:04.659 "num_base_bdevs_discovered": 2, 00:21:04.659 "num_base_bdevs_operational": 2, 00:21:04.659 "base_bdevs_list": [ 00:21:04.659 { 00:21:04.659 "name": "pt1", 00:21:04.659 "uuid": "00000000-0000-0000-0000-000000000001", 00:21:04.659 "is_configured": true, 00:21:04.659 "data_offset": 2048, 00:21:04.659 "data_size": 63488 00:21:04.659 }, 00:21:04.659 { 00:21:04.659 "name": "pt2", 00:21:04.659 "uuid": "00000000-0000-0000-0000-000000000002", 00:21:04.659 "is_configured": true, 00:21:04.659 "data_offset": 2048, 00:21:04.659 "data_size": 63488 00:21:04.659 } 00:21:04.659 ] 00:21:04.659 }' 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:04.659 20:19:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:05.225 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:21:05.225 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:21:05.225 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:05.226 [2024-10-01 20:20:00.291161] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:21:05.226 "name": "raid_bdev1", 00:21:05.226 "aliases": [ 00:21:05.226 "8b33d87d-9289-438d-9414-70be971e2f32" 00:21:05.226 ], 00:21:05.226 "product_name": "Raid Volume", 00:21:05.226 "block_size": 512, 00:21:05.226 "num_blocks": 63488, 00:21:05.226 "uuid": "8b33d87d-9289-438d-9414-70be971e2f32", 00:21:05.226 "assigned_rate_limits": { 00:21:05.226 "rw_ios_per_sec": 0, 00:21:05.226 "rw_mbytes_per_sec": 0, 00:21:05.226 "r_mbytes_per_sec": 0, 00:21:05.226 "w_mbytes_per_sec": 0 00:21:05.226 }, 00:21:05.226 "claimed": false, 00:21:05.226 "zoned": false, 00:21:05.226 "supported_io_types": { 00:21:05.226 "read": true, 00:21:05.226 "write": true, 00:21:05.226 "unmap": false, 00:21:05.226 "flush": false, 00:21:05.226 "reset": true, 00:21:05.226 "nvme_admin": false, 00:21:05.226 "nvme_io": false, 00:21:05.226 "nvme_io_md": false, 00:21:05.226 "write_zeroes": true, 00:21:05.226 "zcopy": false, 00:21:05.226 "get_zone_info": false, 00:21:05.226 "zone_management": false, 00:21:05.226 "zone_append": false, 00:21:05.226 "compare": false, 00:21:05.226 "compare_and_write": false, 00:21:05.226 "abort": false, 00:21:05.226 "seek_hole": false, 00:21:05.226 "seek_data": false, 00:21:05.226 "copy": false, 00:21:05.226 "nvme_iov_md": false 00:21:05.226 }, 00:21:05.226 "memory_domains": [ 00:21:05.226 { 00:21:05.226 "dma_device_id": "system", 00:21:05.226 "dma_device_type": 1 00:21:05.226 }, 00:21:05.226 { 00:21:05.226 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:05.226 "dma_device_type": 2 00:21:05.226 }, 00:21:05.226 { 00:21:05.226 "dma_device_id": "system", 00:21:05.226 "dma_device_type": 1 00:21:05.226 }, 00:21:05.226 { 00:21:05.226 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:05.226 "dma_device_type": 2 00:21:05.226 } 00:21:05.226 ], 00:21:05.226 "driver_specific": { 00:21:05.226 "raid": { 00:21:05.226 "uuid": "8b33d87d-9289-438d-9414-70be971e2f32", 00:21:05.226 "strip_size_kb": 0, 00:21:05.226 "state": "online", 00:21:05.226 "raid_level": "raid1", 00:21:05.226 "superblock": true, 00:21:05.226 "num_base_bdevs": 2, 00:21:05.226 "num_base_bdevs_discovered": 2, 00:21:05.226 "num_base_bdevs_operational": 2, 00:21:05.226 "base_bdevs_list": [ 00:21:05.226 { 00:21:05.226 "name": "pt1", 00:21:05.226 "uuid": "00000000-0000-0000-0000-000000000001", 00:21:05.226 "is_configured": true, 00:21:05.226 "data_offset": 2048, 00:21:05.226 "data_size": 63488 00:21:05.226 }, 00:21:05.226 { 00:21:05.226 "name": "pt2", 00:21:05.226 "uuid": "00000000-0000-0000-0000-000000000002", 00:21:05.226 "is_configured": true, 00:21:05.226 "data_offset": 2048, 00:21:05.226 "data_size": 63488 00:21:05.226 } 00:21:05.226 ] 00:21:05.226 } 00:21:05.226 } 00:21:05.226 }' 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:21:05.226 pt2' 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:05.226 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:05.485 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:05.485 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:05.485 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:05.486 [2024-10-01 20:20:00.563231] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 8b33d87d-9289-438d-9414-70be971e2f32 '!=' 8b33d87d-9289-438d-9414-70be971e2f32 ']' 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:05.486 [2024-10-01 20:20:00.614911] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:05.486 "name": "raid_bdev1", 00:21:05.486 "uuid": "8b33d87d-9289-438d-9414-70be971e2f32", 00:21:05.486 "strip_size_kb": 0, 00:21:05.486 "state": "online", 00:21:05.486 "raid_level": "raid1", 00:21:05.486 "superblock": true, 00:21:05.486 "num_base_bdevs": 2, 00:21:05.486 "num_base_bdevs_discovered": 1, 00:21:05.486 "num_base_bdevs_operational": 1, 00:21:05.486 "base_bdevs_list": [ 00:21:05.486 { 00:21:05.486 "name": null, 00:21:05.486 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:05.486 "is_configured": false, 00:21:05.486 "data_offset": 0, 00:21:05.486 "data_size": 63488 00:21:05.486 }, 00:21:05.486 { 00:21:05.486 "name": "pt2", 00:21:05.486 "uuid": "00000000-0000-0000-0000-000000000002", 00:21:05.486 "is_configured": true, 00:21:05.486 "data_offset": 2048, 00:21:05.486 "data_size": 63488 00:21:05.486 } 00:21:05.486 ] 00:21:05.486 }' 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:05.486 20:20:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:06.055 [2024-10-01 20:20:01.151068] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:21:06.055 [2024-10-01 20:20:01.151120] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:21:06.055 [2024-10-01 20:20:01.151254] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:21:06.055 [2024-10-01 20:20:01.151335] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:21:06.055 [2024-10-01 20:20:01.151355] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=1 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:06.055 [2024-10-01 20:20:01.231070] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:21:06.055 [2024-10-01 20:20:01.231195] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:06.055 [2024-10-01 20:20:01.231222] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:21:06.055 [2024-10-01 20:20:01.231241] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:06.055 [2024-10-01 20:20:01.234528] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:06.055 [2024-10-01 20:20:01.234589] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:21:06.055 [2024-10-01 20:20:01.234704] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:21:06.055 [2024-10-01 20:20:01.234803] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:21:06.055 [2024-10-01 20:20:01.234961] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:21:06.055 [2024-10-01 20:20:01.234986] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:21:06.055 [2024-10-01 20:20:01.235292] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:21:06.055 [2024-10-01 20:20:01.235512] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:21:06.055 [2024-10-01 20:20:01.235538] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:21:06.055 [2024-10-01 20:20:01.235859] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:06.055 pt2 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:06.055 "name": "raid_bdev1", 00:21:06.055 "uuid": "8b33d87d-9289-438d-9414-70be971e2f32", 00:21:06.055 "strip_size_kb": 0, 00:21:06.055 "state": "online", 00:21:06.055 "raid_level": "raid1", 00:21:06.055 "superblock": true, 00:21:06.055 "num_base_bdevs": 2, 00:21:06.055 "num_base_bdevs_discovered": 1, 00:21:06.055 "num_base_bdevs_operational": 1, 00:21:06.055 "base_bdevs_list": [ 00:21:06.055 { 00:21:06.055 "name": null, 00:21:06.055 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:06.055 "is_configured": false, 00:21:06.055 "data_offset": 2048, 00:21:06.055 "data_size": 63488 00:21:06.055 }, 00:21:06.055 { 00:21:06.055 "name": "pt2", 00:21:06.055 "uuid": "00000000-0000-0000-0000-000000000002", 00:21:06.055 "is_configured": true, 00:21:06.055 "data_offset": 2048, 00:21:06.055 "data_size": 63488 00:21:06.055 } 00:21:06.055 ] 00:21:06.055 }' 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:06.055 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:06.744 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:21:06.744 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:06.744 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:06.744 [2024-10-01 20:20:01.771364] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:21:06.744 [2024-10-01 20:20:01.771549] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:21:06.744 [2024-10-01 20:20:01.771791] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:21:06.744 [2024-10-01 20:20:01.771986] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:21:06.744 [2024-10-01 20:20:01.772132] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:21:06.744 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:06.744 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:06.744 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:21:06.744 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:06.744 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:06.744 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:06.744 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:21:06.744 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:21:06.744 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:06.745 [2024-10-01 20:20:01.835350] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:21:06.745 [2024-10-01 20:20:01.835434] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:06.745 [2024-10-01 20:20:01.835465] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:21:06.745 [2024-10-01 20:20:01.835480] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:06.745 [2024-10-01 20:20:01.838773] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:06.745 [2024-10-01 20:20:01.838836] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:21:06.745 [2024-10-01 20:20:01.838944] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:21:06.745 [2024-10-01 20:20:01.839000] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:21:06.745 [2024-10-01 20:20:01.839176] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:21:06.745 [2024-10-01 20:20:01.839211] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:21:06.745 [2024-10-01 20:20:01.839235] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:21:06.745 [2024-10-01 20:20:01.839327] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:21:06.745 [2024-10-01 20:20:01.839464] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:21:06.745 [2024-10-01 20:20:01.839481] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:21:06.745 [2024-10-01 20:20:01.839837] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:21:06.745 [2024-10-01 20:20:01.840028] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:21:06.745 [2024-10-01 20:20:01.840048] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:21:06.745 pt1 00:21:06.745 [2024-10-01 20:20:01.840274] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:06.745 "name": "raid_bdev1", 00:21:06.745 "uuid": "8b33d87d-9289-438d-9414-70be971e2f32", 00:21:06.745 "strip_size_kb": 0, 00:21:06.745 "state": "online", 00:21:06.745 "raid_level": "raid1", 00:21:06.745 "superblock": true, 00:21:06.745 "num_base_bdevs": 2, 00:21:06.745 "num_base_bdevs_discovered": 1, 00:21:06.745 "num_base_bdevs_operational": 1, 00:21:06.745 "base_bdevs_list": [ 00:21:06.745 { 00:21:06.745 "name": null, 00:21:06.745 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:06.745 "is_configured": false, 00:21:06.745 "data_offset": 2048, 00:21:06.745 "data_size": 63488 00:21:06.745 }, 00:21:06.745 { 00:21:06.745 "name": "pt2", 00:21:06.745 "uuid": "00000000-0000-0000-0000-000000000002", 00:21:06.745 "is_configured": true, 00:21:06.745 "data_offset": 2048, 00:21:06.745 "data_size": 63488 00:21:06.745 } 00:21:06.745 ] 00:21:06.745 }' 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:06.745 20:20:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:07.313 [2024-10-01 20:20:02.432125] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 8b33d87d-9289-438d-9414-70be971e2f32 '!=' 8b33d87d-9289-438d-9414-70be971e2f32 ']' 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 63659 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 63659 ']' 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 63659 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 63659 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:21:07.313 killing process with pid 63659 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 63659' 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 63659 00:21:07.313 [2024-10-01 20:20:02.507753] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:21:07.313 20:20:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 63659 00:21:07.313 [2024-10-01 20:20:02.507883] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:21:07.313 [2024-10-01 20:20:02.507951] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:21:07.313 [2024-10-01 20:20:02.507974] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:21:07.571 [2024-10-01 20:20:02.688025] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:21:09.474 20:20:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:21:09.474 00:21:09.474 real 0m7.562s 00:21:09.474 user 0m11.442s 00:21:09.474 sys 0m1.092s 00:21:09.474 20:20:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:21:09.474 ************************************ 00:21:09.474 END TEST raid_superblock_test 00:21:09.474 ************************************ 00:21:09.474 20:20:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:09.474 20:20:04 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 2 read 00:21:09.474 20:20:04 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:21:09.474 20:20:04 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:21:09.474 20:20:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:21:09.474 ************************************ 00:21:09.474 START TEST raid_read_error_test 00:21:09.474 ************************************ 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 2 read 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.cZRICbAtMa 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=64001 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 64001 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 64001 ']' 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:21:09.474 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:09.474 20:20:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:09.475 20:20:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:21:09.475 20:20:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:09.475 [2024-10-01 20:20:04.671879] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:21:09.475 [2024-10-01 20:20:04.672077] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid64001 ] 00:21:09.733 [2024-10-01 20:20:04.849662] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:09.992 [2024-10-01 20:20:05.096649] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:21:10.250 [2024-10-01 20:20:05.302396] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:10.250 [2024-10-01 20:20:05.302511] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:10.509 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:21:10.509 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:21:10.509 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:21:10.509 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:21:10.509 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:10.509 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:10.509 BaseBdev1_malloc 00:21:10.509 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:10.509 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:21:10.509 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:10.509 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:10.769 true 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:10.769 [2024-10-01 20:20:05.768636] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:21:10.769 [2024-10-01 20:20:05.768699] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:10.769 [2024-10-01 20:20:05.768770] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:21:10.769 [2024-10-01 20:20:05.768795] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:10.769 [2024-10-01 20:20:05.771769] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:10.769 [2024-10-01 20:20:05.771824] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:21:10.769 BaseBdev1 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:10.769 BaseBdev2_malloc 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:10.769 true 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:10.769 [2024-10-01 20:20:05.829952] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:21:10.769 [2024-10-01 20:20:05.830024] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:10.769 [2024-10-01 20:20:05.830051] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:21:10.769 [2024-10-01 20:20:05.830071] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:10.769 [2024-10-01 20:20:05.833182] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:10.769 [2024-10-01 20:20:05.833276] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:21:10.769 BaseBdev2 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:10.769 [2024-10-01 20:20:05.838181] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:21:10.769 [2024-10-01 20:20:05.840770] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:21:10.769 [2024-10-01 20:20:05.841110] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:21:10.769 [2024-10-01 20:20:05.841134] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:21:10.769 [2024-10-01 20:20:05.841435] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:21:10.769 [2024-10-01 20:20:05.841663] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:21:10.769 [2024-10-01 20:20:05.841680] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:21:10.769 [2024-10-01 20:20:05.841893] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:10.769 "name": "raid_bdev1", 00:21:10.769 "uuid": "c93be7f9-a9f5-4aca-8d1f-9a379234eeac", 00:21:10.769 "strip_size_kb": 0, 00:21:10.769 "state": "online", 00:21:10.769 "raid_level": "raid1", 00:21:10.769 "superblock": true, 00:21:10.769 "num_base_bdevs": 2, 00:21:10.769 "num_base_bdevs_discovered": 2, 00:21:10.769 "num_base_bdevs_operational": 2, 00:21:10.769 "base_bdevs_list": [ 00:21:10.769 { 00:21:10.769 "name": "BaseBdev1", 00:21:10.769 "uuid": "3df46a8c-9529-53f2-ab16-f9f53df3e850", 00:21:10.769 "is_configured": true, 00:21:10.769 "data_offset": 2048, 00:21:10.769 "data_size": 63488 00:21:10.769 }, 00:21:10.769 { 00:21:10.769 "name": "BaseBdev2", 00:21:10.769 "uuid": "1e4ee7b8-1914-5f35-8b1e-51539036cb37", 00:21:10.769 "is_configured": true, 00:21:10.769 "data_offset": 2048, 00:21:10.769 "data_size": 63488 00:21:10.769 } 00:21:10.769 ] 00:21:10.769 }' 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:10.769 20:20:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:11.337 20:20:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:21:11.337 20:20:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:21:11.337 [2024-10-01 20:20:06.491759] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:12.275 "name": "raid_bdev1", 00:21:12.275 "uuid": "c93be7f9-a9f5-4aca-8d1f-9a379234eeac", 00:21:12.275 "strip_size_kb": 0, 00:21:12.275 "state": "online", 00:21:12.275 "raid_level": "raid1", 00:21:12.275 "superblock": true, 00:21:12.275 "num_base_bdevs": 2, 00:21:12.275 "num_base_bdevs_discovered": 2, 00:21:12.275 "num_base_bdevs_operational": 2, 00:21:12.275 "base_bdevs_list": [ 00:21:12.275 { 00:21:12.275 "name": "BaseBdev1", 00:21:12.275 "uuid": "3df46a8c-9529-53f2-ab16-f9f53df3e850", 00:21:12.275 "is_configured": true, 00:21:12.275 "data_offset": 2048, 00:21:12.275 "data_size": 63488 00:21:12.275 }, 00:21:12.275 { 00:21:12.275 "name": "BaseBdev2", 00:21:12.275 "uuid": "1e4ee7b8-1914-5f35-8b1e-51539036cb37", 00:21:12.275 "is_configured": true, 00:21:12.275 "data_offset": 2048, 00:21:12.275 "data_size": 63488 00:21:12.275 } 00:21:12.275 ] 00:21:12.275 }' 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:12.275 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:12.842 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:21:12.842 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:12.842 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:12.842 [2024-10-01 20:20:07.922914] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:21:12.842 [2024-10-01 20:20:07.923119] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:21:12.842 [2024-10-01 20:20:07.926602] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:21:12.842 { 00:21:12.842 "results": [ 00:21:12.842 { 00:21:12.842 "job": "raid_bdev1", 00:21:12.842 "core_mask": "0x1", 00:21:12.842 "workload": "randrw", 00:21:12.842 "percentage": 50, 00:21:12.842 "status": "finished", 00:21:12.842 "queue_depth": 1, 00:21:12.842 "io_size": 131072, 00:21:12.842 "runtime": 1.428515, 00:21:12.842 "iops": 11737.363625863221, 00:21:12.842 "mibps": 1467.1704532329027, 00:21:12.842 "io_failed": 0, 00:21:12.842 "io_timeout": 0, 00:21:12.842 "avg_latency_us": 80.84620244311066, 00:21:12.842 "min_latency_us": 39.79636363636364, 00:21:12.842 "max_latency_us": 1832.0290909090909 00:21:12.842 } 00:21:12.842 ], 00:21:12.842 "core_count": 1 00:21:12.842 } 00:21:12.842 [2024-10-01 20:20:07.926855] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:12.842 [2024-10-01 20:20:07.927041] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:21:12.842 [2024-10-01 20:20:07.927072] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:21:12.842 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:12.842 20:20:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 64001 00:21:12.842 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 64001 ']' 00:21:12.842 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 64001 00:21:12.842 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:21:12.842 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:21:12.842 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 64001 00:21:12.842 killing process with pid 64001 00:21:12.842 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:21:12.842 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:21:12.842 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 64001' 00:21:12.842 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 64001 00:21:12.842 [2024-10-01 20:20:07.971380] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:21:12.842 20:20:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 64001 00:21:13.101 [2024-10-01 20:20:08.101367] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:21:15.004 20:20:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.cZRICbAtMa 00:21:15.004 20:20:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:21:15.004 20:20:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:21:15.004 20:20:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:21:15.004 20:20:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:21:15.004 20:20:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:21:15.004 20:20:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:21:15.004 20:20:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:21:15.004 00:21:15.004 real 0m5.443s 00:21:15.004 user 0m6.392s 00:21:15.004 sys 0m0.709s 00:21:15.004 20:20:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:21:15.004 20:20:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:15.004 ************************************ 00:21:15.004 END TEST raid_read_error_test 00:21:15.004 ************************************ 00:21:15.004 20:20:10 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 2 write 00:21:15.004 20:20:10 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:21:15.004 20:20:10 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:21:15.004 20:20:10 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:21:15.004 ************************************ 00:21:15.004 START TEST raid_write_error_test 00:21:15.004 ************************************ 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 2 write 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.BJDM6SKM4Q 00:21:15.004 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=64158 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 64158 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 64158 ']' 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:21:15.004 20:20:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:15.004 [2024-10-01 20:20:10.179369] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:21:15.004 [2024-10-01 20:20:10.179589] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid64158 ] 00:21:15.263 [2024-10-01 20:20:10.363120] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:15.522 [2024-10-01 20:20:10.629132] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:21:15.781 [2024-10-01 20:20:10.850607] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:15.781 [2024-10-01 20:20:10.850673] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:16.039 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:21:16.039 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:21:16.039 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:21:16.039 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:21:16.039 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:16.039 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:16.298 BaseBdev1_malloc 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:16.298 true 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:16.298 [2024-10-01 20:20:11.349018] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:21:16.298 [2024-10-01 20:20:11.349092] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:16.298 [2024-10-01 20:20:11.349120] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:21:16.298 [2024-10-01 20:20:11.349140] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:16.298 [2024-10-01 20:20:11.352213] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:16.298 [2024-10-01 20:20:11.352290] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:21:16.298 BaseBdev1 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:16.298 BaseBdev2_malloc 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:16.298 true 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:16.298 [2024-10-01 20:20:11.415618] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:21:16.298 [2024-10-01 20:20:11.415740] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:16.298 [2024-10-01 20:20:11.415773] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:21:16.298 [2024-10-01 20:20:11.415793] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:16.298 [2024-10-01 20:20:11.418922] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:16.298 [2024-10-01 20:20:11.419019] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:21:16.298 BaseBdev2 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:16.298 [2024-10-01 20:20:11.423762] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:21:16.298 [2024-10-01 20:20:11.426654] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:21:16.298 [2024-10-01 20:20:11.427073] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:21:16.298 [2024-10-01 20:20:11.427263] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:21:16.298 [2024-10-01 20:20:11.427641] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:21:16.298 [2024-10-01 20:20:11.428010] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:21:16.298 [2024-10-01 20:20:11.428160] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:21:16.298 [2024-10-01 20:20:11.428612] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:16.298 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:16.299 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:21:16.299 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:16.299 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:16.299 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:16.299 "name": "raid_bdev1", 00:21:16.299 "uuid": "a0638e24-359e-43c5-9c44-721fc4594e95", 00:21:16.299 "strip_size_kb": 0, 00:21:16.299 "state": "online", 00:21:16.299 "raid_level": "raid1", 00:21:16.299 "superblock": true, 00:21:16.299 "num_base_bdevs": 2, 00:21:16.299 "num_base_bdevs_discovered": 2, 00:21:16.299 "num_base_bdevs_operational": 2, 00:21:16.299 "base_bdevs_list": [ 00:21:16.299 { 00:21:16.299 "name": "BaseBdev1", 00:21:16.299 "uuid": "1fb3145b-1c03-5d13-8afc-97847a422ba8", 00:21:16.299 "is_configured": true, 00:21:16.299 "data_offset": 2048, 00:21:16.299 "data_size": 63488 00:21:16.299 }, 00:21:16.299 { 00:21:16.299 "name": "BaseBdev2", 00:21:16.299 "uuid": "3889a2b1-2bcf-5316-a79c-3eeccc543635", 00:21:16.299 "is_configured": true, 00:21:16.299 "data_offset": 2048, 00:21:16.299 "data_size": 63488 00:21:16.299 } 00:21:16.299 ] 00:21:16.299 }' 00:21:16.299 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:16.299 20:20:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:16.866 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:21:16.866 20:20:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:21:16.866 [2024-10-01 20:20:12.086239] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:17.802 [2024-10-01 20:20:12.963074] bdev_raid.c:2272:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:21:17.802 [2024-10-01 20:20:12.963184] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:21:17.802 [2024-10-01 20:20:12.963408] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000005ee0 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=1 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:17.802 20:20:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:17.802 20:20:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:17.802 "name": "raid_bdev1", 00:21:17.802 "uuid": "a0638e24-359e-43c5-9c44-721fc4594e95", 00:21:17.802 "strip_size_kb": 0, 00:21:17.802 "state": "online", 00:21:17.802 "raid_level": "raid1", 00:21:17.802 "superblock": true, 00:21:17.802 "num_base_bdevs": 2, 00:21:17.802 "num_base_bdevs_discovered": 1, 00:21:17.802 "num_base_bdevs_operational": 1, 00:21:17.802 "base_bdevs_list": [ 00:21:17.802 { 00:21:17.802 "name": null, 00:21:17.802 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:17.802 "is_configured": false, 00:21:17.802 "data_offset": 0, 00:21:17.802 "data_size": 63488 00:21:17.802 }, 00:21:17.802 { 00:21:17.802 "name": "BaseBdev2", 00:21:17.802 "uuid": "3889a2b1-2bcf-5316-a79c-3eeccc543635", 00:21:17.802 "is_configured": true, 00:21:17.802 "data_offset": 2048, 00:21:17.802 "data_size": 63488 00:21:17.802 } 00:21:17.802 ] 00:21:17.802 }' 00:21:17.802 20:20:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:17.802 20:20:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:18.405 20:20:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:21:18.405 20:20:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:18.405 20:20:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:18.405 [2024-10-01 20:20:13.506619] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:21:18.405 [2024-10-01 20:20:13.506656] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:21:18.405 [2024-10-01 20:20:13.510006] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:21:18.405 [2024-10-01 20:20:13.510201] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:18.405 [2024-10-01 20:20:13.510298] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:21:18.405 [2024-10-01 20:20:13.510316] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:21:18.405 { 00:21:18.405 "results": [ 00:21:18.405 { 00:21:18.405 "job": "raid_bdev1", 00:21:18.405 "core_mask": "0x1", 00:21:18.405 "workload": "randrw", 00:21:18.405 "percentage": 50, 00:21:18.405 "status": "finished", 00:21:18.405 "queue_depth": 1, 00:21:18.405 "io_size": 131072, 00:21:18.405 "runtime": 1.417777, 00:21:18.405 "iops": 13808.941744717258, 00:21:18.405 "mibps": 1726.1177180896573, 00:21:18.405 "io_failed": 0, 00:21:18.405 "io_timeout": 0, 00:21:18.405 "avg_latency_us": 68.009309150345, 00:21:18.405 "min_latency_us": 38.63272727272727, 00:21:18.405 "max_latency_us": 1824.581818181818 00:21:18.405 } 00:21:18.405 ], 00:21:18.405 "core_count": 1 00:21:18.405 } 00:21:18.405 20:20:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:18.405 20:20:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 64158 00:21:18.405 20:20:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 64158 ']' 00:21:18.405 20:20:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 64158 00:21:18.405 20:20:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:21:18.405 20:20:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:21:18.405 20:20:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 64158 00:21:18.405 killing process with pid 64158 00:21:18.405 20:20:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:21:18.405 20:20:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:21:18.405 20:20:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 64158' 00:21:18.405 20:20:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 64158 00:21:18.405 [2024-10-01 20:20:13.550657] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:21:18.405 20:20:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 64158 00:21:18.665 [2024-10-01 20:20:13.676324] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:21:20.569 20:20:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.BJDM6SKM4Q 00:21:20.569 20:20:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:21:20.569 20:20:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:21:20.569 20:20:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:21:20.569 20:20:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:21:20.569 20:20:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:21:20.569 20:20:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:21:20.569 20:20:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:21:20.569 00:21:20.569 real 0m5.533s 00:21:20.569 user 0m6.559s 00:21:20.569 sys 0m0.718s 00:21:20.569 20:20:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:21:20.569 20:20:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:20.569 ************************************ 00:21:20.569 END TEST raid_write_error_test 00:21:20.569 ************************************ 00:21:20.569 20:20:15 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:21:20.569 20:20:15 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:21:20.569 20:20:15 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 3 false 00:21:20.569 20:20:15 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:21:20.569 20:20:15 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:21:20.569 20:20:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:21:20.569 ************************************ 00:21:20.569 START TEST raid_state_function_test 00:21:20.569 ************************************ 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 3 false 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=64307 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 64307' 00:21:20.569 Process raid pid: 64307 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 64307 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 64307 ']' 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:20.569 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:21:20.569 20:20:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:20.569 [2024-10-01 20:20:15.758690] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:21:20.569 [2024-10-01 20:20:15.758892] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:20.828 [2024-10-01 20:20:15.939406] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:21.087 [2024-10-01 20:20:16.247530] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:21:21.344 [2024-10-01 20:20:16.462102] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:21.344 [2024-10-01 20:20:16.462165] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:21.911 [2024-10-01 20:20:16.900442] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:21:21.911 [2024-10-01 20:20:16.900724] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:21:21.911 [2024-10-01 20:20:16.900860] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:21:21.911 [2024-10-01 20:20:16.900908] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:21:21.911 [2024-10-01 20:20:16.900923] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:21:21.911 [2024-10-01 20:20:16.900942] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:21.911 20:20:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:21.912 20:20:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:21.912 20:20:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:21.912 20:20:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:21.912 20:20:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:21.912 20:20:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:21.912 "name": "Existed_Raid", 00:21:21.912 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:21.912 "strip_size_kb": 64, 00:21:21.912 "state": "configuring", 00:21:21.912 "raid_level": "raid0", 00:21:21.912 "superblock": false, 00:21:21.912 "num_base_bdevs": 3, 00:21:21.912 "num_base_bdevs_discovered": 0, 00:21:21.912 "num_base_bdevs_operational": 3, 00:21:21.912 "base_bdevs_list": [ 00:21:21.912 { 00:21:21.912 "name": "BaseBdev1", 00:21:21.912 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:21.912 "is_configured": false, 00:21:21.912 "data_offset": 0, 00:21:21.912 "data_size": 0 00:21:21.912 }, 00:21:21.912 { 00:21:21.912 "name": "BaseBdev2", 00:21:21.912 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:21.912 "is_configured": false, 00:21:21.912 "data_offset": 0, 00:21:21.912 "data_size": 0 00:21:21.912 }, 00:21:21.912 { 00:21:21.912 "name": "BaseBdev3", 00:21:21.912 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:21.912 "is_configured": false, 00:21:21.912 "data_offset": 0, 00:21:21.912 "data_size": 0 00:21:21.912 } 00:21:21.912 ] 00:21:21.912 }' 00:21:21.912 20:20:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:21.912 20:20:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:22.480 [2024-10-01 20:20:17.432519] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:21:22.480 [2024-10-01 20:20:17.432604] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:22.480 [2024-10-01 20:20:17.440513] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:21:22.480 [2024-10-01 20:20:17.440707] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:21:22.480 [2024-10-01 20:20:17.440878] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:21:22.480 [2024-10-01 20:20:17.440960] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:21:22.480 [2024-10-01 20:20:17.441207] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:21:22.480 [2024-10-01 20:20:17.441269] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:22.480 [2024-10-01 20:20:17.486227] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:21:22.480 BaseBdev1 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:22.480 [ 00:21:22.480 { 00:21:22.480 "name": "BaseBdev1", 00:21:22.480 "aliases": [ 00:21:22.480 "37802355-2dcb-44f0-9d25-a4e7cfa4880f" 00:21:22.480 ], 00:21:22.480 "product_name": "Malloc disk", 00:21:22.480 "block_size": 512, 00:21:22.480 "num_blocks": 65536, 00:21:22.480 "uuid": "37802355-2dcb-44f0-9d25-a4e7cfa4880f", 00:21:22.480 "assigned_rate_limits": { 00:21:22.480 "rw_ios_per_sec": 0, 00:21:22.480 "rw_mbytes_per_sec": 0, 00:21:22.480 "r_mbytes_per_sec": 0, 00:21:22.480 "w_mbytes_per_sec": 0 00:21:22.480 }, 00:21:22.480 "claimed": true, 00:21:22.480 "claim_type": "exclusive_write", 00:21:22.480 "zoned": false, 00:21:22.480 "supported_io_types": { 00:21:22.480 "read": true, 00:21:22.480 "write": true, 00:21:22.480 "unmap": true, 00:21:22.480 "flush": true, 00:21:22.480 "reset": true, 00:21:22.480 "nvme_admin": false, 00:21:22.480 "nvme_io": false, 00:21:22.480 "nvme_io_md": false, 00:21:22.480 "write_zeroes": true, 00:21:22.480 "zcopy": true, 00:21:22.480 "get_zone_info": false, 00:21:22.480 "zone_management": false, 00:21:22.480 "zone_append": false, 00:21:22.480 "compare": false, 00:21:22.480 "compare_and_write": false, 00:21:22.480 "abort": true, 00:21:22.480 "seek_hole": false, 00:21:22.480 "seek_data": false, 00:21:22.480 "copy": true, 00:21:22.480 "nvme_iov_md": false 00:21:22.480 }, 00:21:22.480 "memory_domains": [ 00:21:22.480 { 00:21:22.480 "dma_device_id": "system", 00:21:22.480 "dma_device_type": 1 00:21:22.480 }, 00:21:22.480 { 00:21:22.480 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:22.480 "dma_device_type": 2 00:21:22.480 } 00:21:22.480 ], 00:21:22.480 "driver_specific": {} 00:21:22.480 } 00:21:22.480 ] 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:22.480 "name": "Existed_Raid", 00:21:22.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:22.480 "strip_size_kb": 64, 00:21:22.480 "state": "configuring", 00:21:22.480 "raid_level": "raid0", 00:21:22.480 "superblock": false, 00:21:22.480 "num_base_bdevs": 3, 00:21:22.480 "num_base_bdevs_discovered": 1, 00:21:22.480 "num_base_bdevs_operational": 3, 00:21:22.480 "base_bdevs_list": [ 00:21:22.480 { 00:21:22.480 "name": "BaseBdev1", 00:21:22.480 "uuid": "37802355-2dcb-44f0-9d25-a4e7cfa4880f", 00:21:22.480 "is_configured": true, 00:21:22.480 "data_offset": 0, 00:21:22.480 "data_size": 65536 00:21:22.480 }, 00:21:22.480 { 00:21:22.480 "name": "BaseBdev2", 00:21:22.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:22.480 "is_configured": false, 00:21:22.480 "data_offset": 0, 00:21:22.480 "data_size": 0 00:21:22.480 }, 00:21:22.480 { 00:21:22.480 "name": "BaseBdev3", 00:21:22.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:22.480 "is_configured": false, 00:21:22.480 "data_offset": 0, 00:21:22.480 "data_size": 0 00:21:22.480 } 00:21:22.480 ] 00:21:22.480 }' 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:22.480 20:20:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:23.104 [2024-10-01 20:20:18.038496] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:21:23.104 [2024-10-01 20:20:18.038569] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:23.104 [2024-10-01 20:20:18.046517] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:21:23.104 [2024-10-01 20:20:18.049244] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:21:23.104 [2024-10-01 20:20:18.049496] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:21:23.104 [2024-10-01 20:20:18.049620] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:21:23.104 [2024-10-01 20:20:18.049682] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:23.104 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:23.104 "name": "Existed_Raid", 00:21:23.104 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:23.104 "strip_size_kb": 64, 00:21:23.104 "state": "configuring", 00:21:23.104 "raid_level": "raid0", 00:21:23.104 "superblock": false, 00:21:23.104 "num_base_bdevs": 3, 00:21:23.104 "num_base_bdevs_discovered": 1, 00:21:23.104 "num_base_bdevs_operational": 3, 00:21:23.104 "base_bdevs_list": [ 00:21:23.104 { 00:21:23.104 "name": "BaseBdev1", 00:21:23.104 "uuid": "37802355-2dcb-44f0-9d25-a4e7cfa4880f", 00:21:23.104 "is_configured": true, 00:21:23.104 "data_offset": 0, 00:21:23.104 "data_size": 65536 00:21:23.104 }, 00:21:23.104 { 00:21:23.104 "name": "BaseBdev2", 00:21:23.104 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:23.104 "is_configured": false, 00:21:23.104 "data_offset": 0, 00:21:23.104 "data_size": 0 00:21:23.105 }, 00:21:23.105 { 00:21:23.105 "name": "BaseBdev3", 00:21:23.105 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:23.105 "is_configured": false, 00:21:23.105 "data_offset": 0, 00:21:23.105 "data_size": 0 00:21:23.105 } 00:21:23.105 ] 00:21:23.105 }' 00:21:23.105 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:23.105 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:23.362 BaseBdev2 00:21:23.362 [2024-10-01 20:20:18.585280] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:23.362 [ 00:21:23.362 { 00:21:23.362 "name": "BaseBdev2", 00:21:23.362 "aliases": [ 00:21:23.362 "7e1122b6-ae21-4129-ad23-75f9ef752c5b" 00:21:23.362 ], 00:21:23.362 "product_name": "Malloc disk", 00:21:23.362 "block_size": 512, 00:21:23.362 "num_blocks": 65536, 00:21:23.362 "uuid": "7e1122b6-ae21-4129-ad23-75f9ef752c5b", 00:21:23.362 "assigned_rate_limits": { 00:21:23.362 "rw_ios_per_sec": 0, 00:21:23.362 "rw_mbytes_per_sec": 0, 00:21:23.362 "r_mbytes_per_sec": 0, 00:21:23.362 "w_mbytes_per_sec": 0 00:21:23.362 }, 00:21:23.362 "claimed": true, 00:21:23.362 "claim_type": "exclusive_write", 00:21:23.362 "zoned": false, 00:21:23.362 "supported_io_types": { 00:21:23.362 "read": true, 00:21:23.362 "write": true, 00:21:23.362 "unmap": true, 00:21:23.362 "flush": true, 00:21:23.362 "reset": true, 00:21:23.362 "nvme_admin": false, 00:21:23.362 "nvme_io": false, 00:21:23.362 "nvme_io_md": false, 00:21:23.362 "write_zeroes": true, 00:21:23.362 "zcopy": true, 00:21:23.362 "get_zone_info": false, 00:21:23.362 "zone_management": false, 00:21:23.362 "zone_append": false, 00:21:23.362 "compare": false, 00:21:23.362 "compare_and_write": false, 00:21:23.362 "abort": true, 00:21:23.362 "seek_hole": false, 00:21:23.362 "seek_data": false, 00:21:23.362 "copy": true, 00:21:23.362 "nvme_iov_md": false 00:21:23.362 }, 00:21:23.362 "memory_domains": [ 00:21:23.362 { 00:21:23.362 "dma_device_id": "system", 00:21:23.362 "dma_device_type": 1 00:21:23.362 }, 00:21:23.362 { 00:21:23.362 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:23.362 "dma_device_type": 2 00:21:23.362 } 00:21:23.362 ], 00:21:23.362 "driver_specific": {} 00:21:23.362 } 00:21:23.362 ] 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:23.362 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:23.620 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:23.620 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:23.620 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:23.620 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:23.620 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:23.620 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:23.620 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:23.620 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:23.620 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:23.620 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:23.620 "name": "Existed_Raid", 00:21:23.620 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:23.620 "strip_size_kb": 64, 00:21:23.620 "state": "configuring", 00:21:23.620 "raid_level": "raid0", 00:21:23.620 "superblock": false, 00:21:23.620 "num_base_bdevs": 3, 00:21:23.620 "num_base_bdevs_discovered": 2, 00:21:23.620 "num_base_bdevs_operational": 3, 00:21:23.620 "base_bdevs_list": [ 00:21:23.620 { 00:21:23.620 "name": "BaseBdev1", 00:21:23.620 "uuid": "37802355-2dcb-44f0-9d25-a4e7cfa4880f", 00:21:23.620 "is_configured": true, 00:21:23.620 "data_offset": 0, 00:21:23.620 "data_size": 65536 00:21:23.620 }, 00:21:23.620 { 00:21:23.620 "name": "BaseBdev2", 00:21:23.620 "uuid": "7e1122b6-ae21-4129-ad23-75f9ef752c5b", 00:21:23.620 "is_configured": true, 00:21:23.620 "data_offset": 0, 00:21:23.620 "data_size": 65536 00:21:23.620 }, 00:21:23.620 { 00:21:23.620 "name": "BaseBdev3", 00:21:23.620 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:23.620 "is_configured": false, 00:21:23.620 "data_offset": 0, 00:21:23.620 "data_size": 0 00:21:23.620 } 00:21:23.620 ] 00:21:23.620 }' 00:21:23.620 20:20:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:23.620 20:20:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:24.185 [2024-10-01 20:20:19.184309] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:21:24.185 BaseBdev3 00:21:24.185 [2024-10-01 20:20:19.184674] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:21:24.185 [2024-10-01 20:20:19.184737] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:21:24.185 [2024-10-01 20:20:19.185106] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:21:24.185 [2024-10-01 20:20:19.185366] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:21:24.185 [2024-10-01 20:20:19.185385] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:21:24.185 [2024-10-01 20:20:19.185717] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:24.185 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:24.185 [ 00:21:24.185 { 00:21:24.185 "name": "BaseBdev3", 00:21:24.185 "aliases": [ 00:21:24.185 "c8fb8a61-6680-4ede-ba9a-152bec707e57" 00:21:24.185 ], 00:21:24.185 "product_name": "Malloc disk", 00:21:24.185 "block_size": 512, 00:21:24.185 "num_blocks": 65536, 00:21:24.185 "uuid": "c8fb8a61-6680-4ede-ba9a-152bec707e57", 00:21:24.185 "assigned_rate_limits": { 00:21:24.185 "rw_ios_per_sec": 0, 00:21:24.185 "rw_mbytes_per_sec": 0, 00:21:24.185 "r_mbytes_per_sec": 0, 00:21:24.185 "w_mbytes_per_sec": 0 00:21:24.185 }, 00:21:24.185 "claimed": true, 00:21:24.185 "claim_type": "exclusive_write", 00:21:24.185 "zoned": false, 00:21:24.185 "supported_io_types": { 00:21:24.185 "read": true, 00:21:24.186 "write": true, 00:21:24.186 "unmap": true, 00:21:24.186 "flush": true, 00:21:24.186 "reset": true, 00:21:24.186 "nvme_admin": false, 00:21:24.186 "nvme_io": false, 00:21:24.186 "nvme_io_md": false, 00:21:24.186 "write_zeroes": true, 00:21:24.186 "zcopy": true, 00:21:24.186 "get_zone_info": false, 00:21:24.186 "zone_management": false, 00:21:24.186 "zone_append": false, 00:21:24.186 "compare": false, 00:21:24.186 "compare_and_write": false, 00:21:24.186 "abort": true, 00:21:24.186 "seek_hole": false, 00:21:24.186 "seek_data": false, 00:21:24.186 "copy": true, 00:21:24.186 "nvme_iov_md": false 00:21:24.186 }, 00:21:24.186 "memory_domains": [ 00:21:24.186 { 00:21:24.186 "dma_device_id": "system", 00:21:24.186 "dma_device_type": 1 00:21:24.186 }, 00:21:24.186 { 00:21:24.186 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:24.186 "dma_device_type": 2 00:21:24.186 } 00:21:24.186 ], 00:21:24.186 "driver_specific": {} 00:21:24.186 } 00:21:24.186 ] 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:24.186 "name": "Existed_Raid", 00:21:24.186 "uuid": "0048f93b-8cd0-456f-9a2f-816cf27d587c", 00:21:24.186 "strip_size_kb": 64, 00:21:24.186 "state": "online", 00:21:24.186 "raid_level": "raid0", 00:21:24.186 "superblock": false, 00:21:24.186 "num_base_bdevs": 3, 00:21:24.186 "num_base_bdevs_discovered": 3, 00:21:24.186 "num_base_bdevs_operational": 3, 00:21:24.186 "base_bdevs_list": [ 00:21:24.186 { 00:21:24.186 "name": "BaseBdev1", 00:21:24.186 "uuid": "37802355-2dcb-44f0-9d25-a4e7cfa4880f", 00:21:24.186 "is_configured": true, 00:21:24.186 "data_offset": 0, 00:21:24.186 "data_size": 65536 00:21:24.186 }, 00:21:24.186 { 00:21:24.186 "name": "BaseBdev2", 00:21:24.186 "uuid": "7e1122b6-ae21-4129-ad23-75f9ef752c5b", 00:21:24.186 "is_configured": true, 00:21:24.186 "data_offset": 0, 00:21:24.186 "data_size": 65536 00:21:24.186 }, 00:21:24.186 { 00:21:24.186 "name": "BaseBdev3", 00:21:24.186 "uuid": "c8fb8a61-6680-4ede-ba9a-152bec707e57", 00:21:24.186 "is_configured": true, 00:21:24.186 "data_offset": 0, 00:21:24.186 "data_size": 65536 00:21:24.186 } 00:21:24.186 ] 00:21:24.186 }' 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:24.186 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:24.752 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:21:24.752 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:21:24.752 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:21:24.752 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:21:24.752 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:21:24.752 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:21:24.752 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:21:24.752 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:24.752 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:24.752 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:21:24.752 [2024-10-01 20:20:19.741055] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:21:24.752 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:24.752 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:21:24.752 "name": "Existed_Raid", 00:21:24.752 "aliases": [ 00:21:24.752 "0048f93b-8cd0-456f-9a2f-816cf27d587c" 00:21:24.752 ], 00:21:24.752 "product_name": "Raid Volume", 00:21:24.752 "block_size": 512, 00:21:24.752 "num_blocks": 196608, 00:21:24.752 "uuid": "0048f93b-8cd0-456f-9a2f-816cf27d587c", 00:21:24.752 "assigned_rate_limits": { 00:21:24.752 "rw_ios_per_sec": 0, 00:21:24.752 "rw_mbytes_per_sec": 0, 00:21:24.752 "r_mbytes_per_sec": 0, 00:21:24.752 "w_mbytes_per_sec": 0 00:21:24.752 }, 00:21:24.752 "claimed": false, 00:21:24.752 "zoned": false, 00:21:24.752 "supported_io_types": { 00:21:24.752 "read": true, 00:21:24.752 "write": true, 00:21:24.752 "unmap": true, 00:21:24.752 "flush": true, 00:21:24.752 "reset": true, 00:21:24.752 "nvme_admin": false, 00:21:24.752 "nvme_io": false, 00:21:24.752 "nvme_io_md": false, 00:21:24.752 "write_zeroes": true, 00:21:24.752 "zcopy": false, 00:21:24.752 "get_zone_info": false, 00:21:24.752 "zone_management": false, 00:21:24.752 "zone_append": false, 00:21:24.752 "compare": false, 00:21:24.752 "compare_and_write": false, 00:21:24.752 "abort": false, 00:21:24.752 "seek_hole": false, 00:21:24.752 "seek_data": false, 00:21:24.752 "copy": false, 00:21:24.752 "nvme_iov_md": false 00:21:24.752 }, 00:21:24.752 "memory_domains": [ 00:21:24.752 { 00:21:24.752 "dma_device_id": "system", 00:21:24.752 "dma_device_type": 1 00:21:24.752 }, 00:21:24.752 { 00:21:24.752 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:24.752 "dma_device_type": 2 00:21:24.752 }, 00:21:24.752 { 00:21:24.752 "dma_device_id": "system", 00:21:24.752 "dma_device_type": 1 00:21:24.752 }, 00:21:24.752 { 00:21:24.752 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:24.752 "dma_device_type": 2 00:21:24.752 }, 00:21:24.752 { 00:21:24.752 "dma_device_id": "system", 00:21:24.752 "dma_device_type": 1 00:21:24.752 }, 00:21:24.752 { 00:21:24.752 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:24.752 "dma_device_type": 2 00:21:24.752 } 00:21:24.752 ], 00:21:24.752 "driver_specific": { 00:21:24.752 "raid": { 00:21:24.752 "uuid": "0048f93b-8cd0-456f-9a2f-816cf27d587c", 00:21:24.752 "strip_size_kb": 64, 00:21:24.752 "state": "online", 00:21:24.752 "raid_level": "raid0", 00:21:24.752 "superblock": false, 00:21:24.752 "num_base_bdevs": 3, 00:21:24.752 "num_base_bdevs_discovered": 3, 00:21:24.752 "num_base_bdevs_operational": 3, 00:21:24.752 "base_bdevs_list": [ 00:21:24.752 { 00:21:24.752 "name": "BaseBdev1", 00:21:24.752 "uuid": "37802355-2dcb-44f0-9d25-a4e7cfa4880f", 00:21:24.752 "is_configured": true, 00:21:24.752 "data_offset": 0, 00:21:24.752 "data_size": 65536 00:21:24.752 }, 00:21:24.752 { 00:21:24.752 "name": "BaseBdev2", 00:21:24.752 "uuid": "7e1122b6-ae21-4129-ad23-75f9ef752c5b", 00:21:24.752 "is_configured": true, 00:21:24.752 "data_offset": 0, 00:21:24.752 "data_size": 65536 00:21:24.752 }, 00:21:24.752 { 00:21:24.752 "name": "BaseBdev3", 00:21:24.752 "uuid": "c8fb8a61-6680-4ede-ba9a-152bec707e57", 00:21:24.752 "is_configured": true, 00:21:24.752 "data_offset": 0, 00:21:24.752 "data_size": 65536 00:21:24.752 } 00:21:24.753 ] 00:21:24.753 } 00:21:24.753 } 00:21:24.753 }' 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:21:24.753 BaseBdev2 00:21:24.753 BaseBdev3' 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:24.753 20:20:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:25.039 [2024-10-01 20:20:20.052765] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:21:25.039 [2024-10-01 20:20:20.052946] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:21:25.039 [2024-10-01 20:20:20.053147] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:25.039 "name": "Existed_Raid", 00:21:25.039 "uuid": "0048f93b-8cd0-456f-9a2f-816cf27d587c", 00:21:25.039 "strip_size_kb": 64, 00:21:25.039 "state": "offline", 00:21:25.039 "raid_level": "raid0", 00:21:25.039 "superblock": false, 00:21:25.039 "num_base_bdevs": 3, 00:21:25.039 "num_base_bdevs_discovered": 2, 00:21:25.039 "num_base_bdevs_operational": 2, 00:21:25.039 "base_bdevs_list": [ 00:21:25.039 { 00:21:25.039 "name": null, 00:21:25.039 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:25.039 "is_configured": false, 00:21:25.039 "data_offset": 0, 00:21:25.039 "data_size": 65536 00:21:25.039 }, 00:21:25.039 { 00:21:25.039 "name": "BaseBdev2", 00:21:25.039 "uuid": "7e1122b6-ae21-4129-ad23-75f9ef752c5b", 00:21:25.039 "is_configured": true, 00:21:25.039 "data_offset": 0, 00:21:25.039 "data_size": 65536 00:21:25.039 }, 00:21:25.039 { 00:21:25.039 "name": "BaseBdev3", 00:21:25.039 "uuid": "c8fb8a61-6680-4ede-ba9a-152bec707e57", 00:21:25.039 "is_configured": true, 00:21:25.039 "data_offset": 0, 00:21:25.039 "data_size": 65536 00:21:25.039 } 00:21:25.039 ] 00:21:25.039 }' 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:25.039 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:25.606 [2024-10-01 20:20:20.721645] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:25.606 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:25.865 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:21:25.865 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:21:25.865 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:21:25.865 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:25.865 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:25.865 [2024-10-01 20:20:20.868589] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:21:25.865 [2024-10-01 20:20:20.868809] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:21:25.865 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:25.865 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:21:25.865 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:21:25.865 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:25.865 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:25.865 20:20:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:21:25.865 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:25.865 20:20:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:25.865 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:25.866 BaseBdev2 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:25.866 [ 00:21:25.866 { 00:21:25.866 "name": "BaseBdev2", 00:21:25.866 "aliases": [ 00:21:25.866 "599cc638-1eb1-4faa-96bb-cd116bbc0119" 00:21:25.866 ], 00:21:25.866 "product_name": "Malloc disk", 00:21:25.866 "block_size": 512, 00:21:25.866 "num_blocks": 65536, 00:21:25.866 "uuid": "599cc638-1eb1-4faa-96bb-cd116bbc0119", 00:21:25.866 "assigned_rate_limits": { 00:21:25.866 "rw_ios_per_sec": 0, 00:21:25.866 "rw_mbytes_per_sec": 0, 00:21:25.866 "r_mbytes_per_sec": 0, 00:21:25.866 "w_mbytes_per_sec": 0 00:21:25.866 }, 00:21:25.866 "claimed": false, 00:21:25.866 "zoned": false, 00:21:25.866 "supported_io_types": { 00:21:25.866 "read": true, 00:21:25.866 "write": true, 00:21:25.866 "unmap": true, 00:21:25.866 "flush": true, 00:21:25.866 "reset": true, 00:21:25.866 "nvme_admin": false, 00:21:25.866 "nvme_io": false, 00:21:25.866 "nvme_io_md": false, 00:21:25.866 "write_zeroes": true, 00:21:25.866 "zcopy": true, 00:21:25.866 "get_zone_info": false, 00:21:25.866 "zone_management": false, 00:21:25.866 "zone_append": false, 00:21:25.866 "compare": false, 00:21:25.866 "compare_and_write": false, 00:21:25.866 "abort": true, 00:21:25.866 "seek_hole": false, 00:21:25.866 "seek_data": false, 00:21:25.866 "copy": true, 00:21:25.866 "nvme_iov_md": false 00:21:25.866 }, 00:21:25.866 "memory_domains": [ 00:21:25.866 { 00:21:25.866 "dma_device_id": "system", 00:21:25.866 "dma_device_type": 1 00:21:25.866 }, 00:21:25.866 { 00:21:25.866 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:25.866 "dma_device_type": 2 00:21:25.866 } 00:21:25.866 ], 00:21:25.866 "driver_specific": {} 00:21:25.866 } 00:21:25.866 ] 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:25.866 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:26.125 BaseBdev3 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:26.125 [ 00:21:26.125 { 00:21:26.125 "name": "BaseBdev3", 00:21:26.125 "aliases": [ 00:21:26.125 "1de04ba4-0769-4569-8949-baa4f604a8b6" 00:21:26.125 ], 00:21:26.125 "product_name": "Malloc disk", 00:21:26.125 "block_size": 512, 00:21:26.125 "num_blocks": 65536, 00:21:26.125 "uuid": "1de04ba4-0769-4569-8949-baa4f604a8b6", 00:21:26.125 "assigned_rate_limits": { 00:21:26.125 "rw_ios_per_sec": 0, 00:21:26.125 "rw_mbytes_per_sec": 0, 00:21:26.125 "r_mbytes_per_sec": 0, 00:21:26.125 "w_mbytes_per_sec": 0 00:21:26.125 }, 00:21:26.125 "claimed": false, 00:21:26.125 "zoned": false, 00:21:26.125 "supported_io_types": { 00:21:26.125 "read": true, 00:21:26.125 "write": true, 00:21:26.125 "unmap": true, 00:21:26.125 "flush": true, 00:21:26.125 "reset": true, 00:21:26.125 "nvme_admin": false, 00:21:26.125 "nvme_io": false, 00:21:26.125 "nvme_io_md": false, 00:21:26.125 "write_zeroes": true, 00:21:26.125 "zcopy": true, 00:21:26.125 "get_zone_info": false, 00:21:26.125 "zone_management": false, 00:21:26.125 "zone_append": false, 00:21:26.125 "compare": false, 00:21:26.125 "compare_and_write": false, 00:21:26.125 "abort": true, 00:21:26.125 "seek_hole": false, 00:21:26.125 "seek_data": false, 00:21:26.125 "copy": true, 00:21:26.125 "nvme_iov_md": false 00:21:26.125 }, 00:21:26.125 "memory_domains": [ 00:21:26.125 { 00:21:26.125 "dma_device_id": "system", 00:21:26.125 "dma_device_type": 1 00:21:26.125 }, 00:21:26.125 { 00:21:26.125 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:26.125 "dma_device_type": 2 00:21:26.125 } 00:21:26.125 ], 00:21:26.125 "driver_specific": {} 00:21:26.125 } 00:21:26.125 ] 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:26.125 [2024-10-01 20:20:21.161405] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:21:26.125 [2024-10-01 20:20:21.161467] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:21:26.125 [2024-10-01 20:20:21.161510] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:21:26.125 [2024-10-01 20:20:21.164131] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:26.125 "name": "Existed_Raid", 00:21:26.125 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:26.125 "strip_size_kb": 64, 00:21:26.125 "state": "configuring", 00:21:26.125 "raid_level": "raid0", 00:21:26.125 "superblock": false, 00:21:26.125 "num_base_bdevs": 3, 00:21:26.125 "num_base_bdevs_discovered": 2, 00:21:26.125 "num_base_bdevs_operational": 3, 00:21:26.125 "base_bdevs_list": [ 00:21:26.125 { 00:21:26.125 "name": "BaseBdev1", 00:21:26.125 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:26.125 "is_configured": false, 00:21:26.125 "data_offset": 0, 00:21:26.125 "data_size": 0 00:21:26.125 }, 00:21:26.125 { 00:21:26.125 "name": "BaseBdev2", 00:21:26.125 "uuid": "599cc638-1eb1-4faa-96bb-cd116bbc0119", 00:21:26.125 "is_configured": true, 00:21:26.125 "data_offset": 0, 00:21:26.125 "data_size": 65536 00:21:26.125 }, 00:21:26.125 { 00:21:26.125 "name": "BaseBdev3", 00:21:26.125 "uuid": "1de04ba4-0769-4569-8949-baa4f604a8b6", 00:21:26.125 "is_configured": true, 00:21:26.125 "data_offset": 0, 00:21:26.125 "data_size": 65536 00:21:26.125 } 00:21:26.125 ] 00:21:26.125 }' 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:26.125 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:26.693 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:21:26.693 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:26.693 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:26.693 [2024-10-01 20:20:21.685482] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:21:26.693 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:26.693 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:26.693 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:26.693 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:26.693 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:26.693 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:26.694 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:26.694 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:26.694 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:26.694 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:26.694 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:26.694 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:26.694 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:26.694 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:26.694 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:26.694 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:26.694 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:26.694 "name": "Existed_Raid", 00:21:26.694 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:26.694 "strip_size_kb": 64, 00:21:26.694 "state": "configuring", 00:21:26.694 "raid_level": "raid0", 00:21:26.694 "superblock": false, 00:21:26.694 "num_base_bdevs": 3, 00:21:26.694 "num_base_bdevs_discovered": 1, 00:21:26.694 "num_base_bdevs_operational": 3, 00:21:26.694 "base_bdevs_list": [ 00:21:26.694 { 00:21:26.694 "name": "BaseBdev1", 00:21:26.694 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:26.694 "is_configured": false, 00:21:26.694 "data_offset": 0, 00:21:26.694 "data_size": 0 00:21:26.694 }, 00:21:26.694 { 00:21:26.694 "name": null, 00:21:26.694 "uuid": "599cc638-1eb1-4faa-96bb-cd116bbc0119", 00:21:26.694 "is_configured": false, 00:21:26.694 "data_offset": 0, 00:21:26.694 "data_size": 65536 00:21:26.694 }, 00:21:26.694 { 00:21:26.694 "name": "BaseBdev3", 00:21:26.694 "uuid": "1de04ba4-0769-4569-8949-baa4f604a8b6", 00:21:26.694 "is_configured": true, 00:21:26.694 "data_offset": 0, 00:21:26.694 "data_size": 65536 00:21:26.694 } 00:21:26.694 ] 00:21:26.694 }' 00:21:26.694 20:20:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:26.694 20:20:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:27.261 [2024-10-01 20:20:22.292409] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:21:27.261 BaseBdev1 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:27.261 [ 00:21:27.261 { 00:21:27.261 "name": "BaseBdev1", 00:21:27.261 "aliases": [ 00:21:27.261 "9910cc73-245b-4382-bf2e-f2bbe1c93a5a" 00:21:27.261 ], 00:21:27.261 "product_name": "Malloc disk", 00:21:27.261 "block_size": 512, 00:21:27.261 "num_blocks": 65536, 00:21:27.261 "uuid": "9910cc73-245b-4382-bf2e-f2bbe1c93a5a", 00:21:27.261 "assigned_rate_limits": { 00:21:27.261 "rw_ios_per_sec": 0, 00:21:27.261 "rw_mbytes_per_sec": 0, 00:21:27.261 "r_mbytes_per_sec": 0, 00:21:27.261 "w_mbytes_per_sec": 0 00:21:27.261 }, 00:21:27.261 "claimed": true, 00:21:27.261 "claim_type": "exclusive_write", 00:21:27.261 "zoned": false, 00:21:27.261 "supported_io_types": { 00:21:27.261 "read": true, 00:21:27.261 "write": true, 00:21:27.261 "unmap": true, 00:21:27.261 "flush": true, 00:21:27.261 "reset": true, 00:21:27.261 "nvme_admin": false, 00:21:27.261 "nvme_io": false, 00:21:27.261 "nvme_io_md": false, 00:21:27.261 "write_zeroes": true, 00:21:27.261 "zcopy": true, 00:21:27.261 "get_zone_info": false, 00:21:27.261 "zone_management": false, 00:21:27.261 "zone_append": false, 00:21:27.261 "compare": false, 00:21:27.261 "compare_and_write": false, 00:21:27.261 "abort": true, 00:21:27.261 "seek_hole": false, 00:21:27.261 "seek_data": false, 00:21:27.261 "copy": true, 00:21:27.261 "nvme_iov_md": false 00:21:27.261 }, 00:21:27.261 "memory_domains": [ 00:21:27.261 { 00:21:27.261 "dma_device_id": "system", 00:21:27.261 "dma_device_type": 1 00:21:27.261 }, 00:21:27.261 { 00:21:27.261 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:27.261 "dma_device_type": 2 00:21:27.261 } 00:21:27.261 ], 00:21:27.261 "driver_specific": {} 00:21:27.261 } 00:21:27.261 ] 00:21:27.261 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:27.262 "name": "Existed_Raid", 00:21:27.262 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:27.262 "strip_size_kb": 64, 00:21:27.262 "state": "configuring", 00:21:27.262 "raid_level": "raid0", 00:21:27.262 "superblock": false, 00:21:27.262 "num_base_bdevs": 3, 00:21:27.262 "num_base_bdevs_discovered": 2, 00:21:27.262 "num_base_bdevs_operational": 3, 00:21:27.262 "base_bdevs_list": [ 00:21:27.262 { 00:21:27.262 "name": "BaseBdev1", 00:21:27.262 "uuid": "9910cc73-245b-4382-bf2e-f2bbe1c93a5a", 00:21:27.262 "is_configured": true, 00:21:27.262 "data_offset": 0, 00:21:27.262 "data_size": 65536 00:21:27.262 }, 00:21:27.262 { 00:21:27.262 "name": null, 00:21:27.262 "uuid": "599cc638-1eb1-4faa-96bb-cd116bbc0119", 00:21:27.262 "is_configured": false, 00:21:27.262 "data_offset": 0, 00:21:27.262 "data_size": 65536 00:21:27.262 }, 00:21:27.262 { 00:21:27.262 "name": "BaseBdev3", 00:21:27.262 "uuid": "1de04ba4-0769-4569-8949-baa4f604a8b6", 00:21:27.262 "is_configured": true, 00:21:27.262 "data_offset": 0, 00:21:27.262 "data_size": 65536 00:21:27.262 } 00:21:27.262 ] 00:21:27.262 }' 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:27.262 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:27.830 [2024-10-01 20:20:22.912767] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:27.830 "name": "Existed_Raid", 00:21:27.830 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:27.830 "strip_size_kb": 64, 00:21:27.830 "state": "configuring", 00:21:27.830 "raid_level": "raid0", 00:21:27.830 "superblock": false, 00:21:27.830 "num_base_bdevs": 3, 00:21:27.830 "num_base_bdevs_discovered": 1, 00:21:27.830 "num_base_bdevs_operational": 3, 00:21:27.830 "base_bdevs_list": [ 00:21:27.830 { 00:21:27.830 "name": "BaseBdev1", 00:21:27.830 "uuid": "9910cc73-245b-4382-bf2e-f2bbe1c93a5a", 00:21:27.830 "is_configured": true, 00:21:27.830 "data_offset": 0, 00:21:27.830 "data_size": 65536 00:21:27.830 }, 00:21:27.830 { 00:21:27.830 "name": null, 00:21:27.830 "uuid": "599cc638-1eb1-4faa-96bb-cd116bbc0119", 00:21:27.830 "is_configured": false, 00:21:27.830 "data_offset": 0, 00:21:27.830 "data_size": 65536 00:21:27.830 }, 00:21:27.830 { 00:21:27.830 "name": null, 00:21:27.830 "uuid": "1de04ba4-0769-4569-8949-baa4f604a8b6", 00:21:27.830 "is_configured": false, 00:21:27.830 "data_offset": 0, 00:21:27.830 "data_size": 65536 00:21:27.830 } 00:21:27.830 ] 00:21:27.830 }' 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:27.830 20:20:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:28.398 [2024-10-01 20:20:23.500905] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:28.398 "name": "Existed_Raid", 00:21:28.398 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:28.398 "strip_size_kb": 64, 00:21:28.398 "state": "configuring", 00:21:28.398 "raid_level": "raid0", 00:21:28.398 "superblock": false, 00:21:28.398 "num_base_bdevs": 3, 00:21:28.398 "num_base_bdevs_discovered": 2, 00:21:28.398 "num_base_bdevs_operational": 3, 00:21:28.398 "base_bdevs_list": [ 00:21:28.398 { 00:21:28.398 "name": "BaseBdev1", 00:21:28.398 "uuid": "9910cc73-245b-4382-bf2e-f2bbe1c93a5a", 00:21:28.398 "is_configured": true, 00:21:28.398 "data_offset": 0, 00:21:28.398 "data_size": 65536 00:21:28.398 }, 00:21:28.398 { 00:21:28.398 "name": null, 00:21:28.398 "uuid": "599cc638-1eb1-4faa-96bb-cd116bbc0119", 00:21:28.398 "is_configured": false, 00:21:28.398 "data_offset": 0, 00:21:28.398 "data_size": 65536 00:21:28.398 }, 00:21:28.398 { 00:21:28.398 "name": "BaseBdev3", 00:21:28.398 "uuid": "1de04ba4-0769-4569-8949-baa4f604a8b6", 00:21:28.398 "is_configured": true, 00:21:28.398 "data_offset": 0, 00:21:28.398 "data_size": 65536 00:21:28.398 } 00:21:28.398 ] 00:21:28.398 }' 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:28.398 20:20:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:28.966 [2024-10-01 20:20:24.081085] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:28.966 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:28.966 "name": "Existed_Raid", 00:21:28.966 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:28.966 "strip_size_kb": 64, 00:21:28.966 "state": "configuring", 00:21:28.966 "raid_level": "raid0", 00:21:28.966 "superblock": false, 00:21:28.966 "num_base_bdevs": 3, 00:21:28.966 "num_base_bdevs_discovered": 1, 00:21:28.966 "num_base_bdevs_operational": 3, 00:21:28.966 "base_bdevs_list": [ 00:21:28.966 { 00:21:28.966 "name": null, 00:21:28.966 "uuid": "9910cc73-245b-4382-bf2e-f2bbe1c93a5a", 00:21:28.966 "is_configured": false, 00:21:28.966 "data_offset": 0, 00:21:28.966 "data_size": 65536 00:21:28.966 }, 00:21:28.966 { 00:21:28.966 "name": null, 00:21:28.966 "uuid": "599cc638-1eb1-4faa-96bb-cd116bbc0119", 00:21:28.967 "is_configured": false, 00:21:28.967 "data_offset": 0, 00:21:28.967 "data_size": 65536 00:21:28.967 }, 00:21:28.967 { 00:21:28.967 "name": "BaseBdev3", 00:21:28.967 "uuid": "1de04ba4-0769-4569-8949-baa4f604a8b6", 00:21:28.967 "is_configured": true, 00:21:28.967 "data_offset": 0, 00:21:28.967 "data_size": 65536 00:21:28.967 } 00:21:28.967 ] 00:21:28.967 }' 00:21:28.967 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:28.967 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:29.534 [2024-10-01 20:20:24.739863] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:29.534 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:29.793 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:29.793 "name": "Existed_Raid", 00:21:29.793 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:29.793 "strip_size_kb": 64, 00:21:29.793 "state": "configuring", 00:21:29.793 "raid_level": "raid0", 00:21:29.793 "superblock": false, 00:21:29.793 "num_base_bdevs": 3, 00:21:29.793 "num_base_bdevs_discovered": 2, 00:21:29.793 "num_base_bdevs_operational": 3, 00:21:29.793 "base_bdevs_list": [ 00:21:29.793 { 00:21:29.793 "name": null, 00:21:29.793 "uuid": "9910cc73-245b-4382-bf2e-f2bbe1c93a5a", 00:21:29.793 "is_configured": false, 00:21:29.793 "data_offset": 0, 00:21:29.793 "data_size": 65536 00:21:29.793 }, 00:21:29.793 { 00:21:29.793 "name": "BaseBdev2", 00:21:29.793 "uuid": "599cc638-1eb1-4faa-96bb-cd116bbc0119", 00:21:29.793 "is_configured": true, 00:21:29.793 "data_offset": 0, 00:21:29.793 "data_size": 65536 00:21:29.793 }, 00:21:29.793 { 00:21:29.793 "name": "BaseBdev3", 00:21:29.793 "uuid": "1de04ba4-0769-4569-8949-baa4f604a8b6", 00:21:29.793 "is_configured": true, 00:21:29.793 "data_offset": 0, 00:21:29.793 "data_size": 65536 00:21:29.793 } 00:21:29.793 ] 00:21:29.793 }' 00:21:29.793 20:20:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:29.793 20:20:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:30.052 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:30.052 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:21:30.052 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:30.052 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:30.052 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 9910cc73-245b-4382-bf2e-f2bbe1c93a5a 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:30.311 [2024-10-01 20:20:25.426853] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:21:30.311 [2024-10-01 20:20:25.427355] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:21:30.311 [2024-10-01 20:20:25.427400] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:21:30.311 [2024-10-01 20:20:25.427817] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:21:30.311 NewBaseBdev 00:21:30.311 [2024-10-01 20:20:25.428019] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:21:30.311 [2024-10-01 20:20:25.428043] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:21:30.311 [2024-10-01 20:20:25.428379] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:30.311 [ 00:21:30.311 { 00:21:30.311 "name": "NewBaseBdev", 00:21:30.311 "aliases": [ 00:21:30.311 "9910cc73-245b-4382-bf2e-f2bbe1c93a5a" 00:21:30.311 ], 00:21:30.311 "product_name": "Malloc disk", 00:21:30.311 "block_size": 512, 00:21:30.311 "num_blocks": 65536, 00:21:30.311 "uuid": "9910cc73-245b-4382-bf2e-f2bbe1c93a5a", 00:21:30.311 "assigned_rate_limits": { 00:21:30.311 "rw_ios_per_sec": 0, 00:21:30.311 "rw_mbytes_per_sec": 0, 00:21:30.311 "r_mbytes_per_sec": 0, 00:21:30.311 "w_mbytes_per_sec": 0 00:21:30.311 }, 00:21:30.311 "claimed": true, 00:21:30.311 "claim_type": "exclusive_write", 00:21:30.311 "zoned": false, 00:21:30.311 "supported_io_types": { 00:21:30.311 "read": true, 00:21:30.311 "write": true, 00:21:30.311 "unmap": true, 00:21:30.311 "flush": true, 00:21:30.311 "reset": true, 00:21:30.311 "nvme_admin": false, 00:21:30.311 "nvme_io": false, 00:21:30.311 "nvme_io_md": false, 00:21:30.311 "write_zeroes": true, 00:21:30.311 "zcopy": true, 00:21:30.311 "get_zone_info": false, 00:21:30.311 "zone_management": false, 00:21:30.311 "zone_append": false, 00:21:30.311 "compare": false, 00:21:30.311 "compare_and_write": false, 00:21:30.311 "abort": true, 00:21:30.311 "seek_hole": false, 00:21:30.311 "seek_data": false, 00:21:30.311 "copy": true, 00:21:30.311 "nvme_iov_md": false 00:21:30.311 }, 00:21:30.311 "memory_domains": [ 00:21:30.311 { 00:21:30.311 "dma_device_id": "system", 00:21:30.311 "dma_device_type": 1 00:21:30.311 }, 00:21:30.311 { 00:21:30.311 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:30.311 "dma_device_type": 2 00:21:30.311 } 00:21:30.311 ], 00:21:30.311 "driver_specific": {} 00:21:30.311 } 00:21:30.311 ] 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:30.311 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:30.311 "name": "Existed_Raid", 00:21:30.311 "uuid": "b4b71643-c1d2-4dd7-880c-cb7452c6aa75", 00:21:30.311 "strip_size_kb": 64, 00:21:30.311 "state": "online", 00:21:30.311 "raid_level": "raid0", 00:21:30.311 "superblock": false, 00:21:30.311 "num_base_bdevs": 3, 00:21:30.311 "num_base_bdevs_discovered": 3, 00:21:30.311 "num_base_bdevs_operational": 3, 00:21:30.311 "base_bdevs_list": [ 00:21:30.311 { 00:21:30.311 "name": "NewBaseBdev", 00:21:30.311 "uuid": "9910cc73-245b-4382-bf2e-f2bbe1c93a5a", 00:21:30.311 "is_configured": true, 00:21:30.311 "data_offset": 0, 00:21:30.311 "data_size": 65536 00:21:30.311 }, 00:21:30.311 { 00:21:30.311 "name": "BaseBdev2", 00:21:30.311 "uuid": "599cc638-1eb1-4faa-96bb-cd116bbc0119", 00:21:30.311 "is_configured": true, 00:21:30.311 "data_offset": 0, 00:21:30.311 "data_size": 65536 00:21:30.311 }, 00:21:30.311 { 00:21:30.312 "name": "BaseBdev3", 00:21:30.312 "uuid": "1de04ba4-0769-4569-8949-baa4f604a8b6", 00:21:30.312 "is_configured": true, 00:21:30.312 "data_offset": 0, 00:21:30.312 "data_size": 65536 00:21:30.312 } 00:21:30.312 ] 00:21:30.312 }' 00:21:30.312 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:30.312 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:30.881 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:21:30.881 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:21:30.881 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:21:30.881 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:21:30.881 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:21:30.881 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:21:30.881 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:21:30.881 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:30.881 20:20:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:30.881 20:20:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:21:30.881 [2024-10-01 20:20:25.995469] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:21:30.881 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:30.881 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:21:30.881 "name": "Existed_Raid", 00:21:30.881 "aliases": [ 00:21:30.881 "b4b71643-c1d2-4dd7-880c-cb7452c6aa75" 00:21:30.881 ], 00:21:30.881 "product_name": "Raid Volume", 00:21:30.881 "block_size": 512, 00:21:30.881 "num_blocks": 196608, 00:21:30.881 "uuid": "b4b71643-c1d2-4dd7-880c-cb7452c6aa75", 00:21:30.881 "assigned_rate_limits": { 00:21:30.881 "rw_ios_per_sec": 0, 00:21:30.881 "rw_mbytes_per_sec": 0, 00:21:30.881 "r_mbytes_per_sec": 0, 00:21:30.881 "w_mbytes_per_sec": 0 00:21:30.881 }, 00:21:30.881 "claimed": false, 00:21:30.881 "zoned": false, 00:21:30.881 "supported_io_types": { 00:21:30.881 "read": true, 00:21:30.881 "write": true, 00:21:30.881 "unmap": true, 00:21:30.881 "flush": true, 00:21:30.881 "reset": true, 00:21:30.881 "nvme_admin": false, 00:21:30.881 "nvme_io": false, 00:21:30.881 "nvme_io_md": false, 00:21:30.881 "write_zeroes": true, 00:21:30.881 "zcopy": false, 00:21:30.881 "get_zone_info": false, 00:21:30.881 "zone_management": false, 00:21:30.881 "zone_append": false, 00:21:30.881 "compare": false, 00:21:30.881 "compare_and_write": false, 00:21:30.881 "abort": false, 00:21:30.881 "seek_hole": false, 00:21:30.881 "seek_data": false, 00:21:30.881 "copy": false, 00:21:30.881 "nvme_iov_md": false 00:21:30.881 }, 00:21:30.881 "memory_domains": [ 00:21:30.881 { 00:21:30.881 "dma_device_id": "system", 00:21:30.881 "dma_device_type": 1 00:21:30.881 }, 00:21:30.881 { 00:21:30.881 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:30.881 "dma_device_type": 2 00:21:30.881 }, 00:21:30.881 { 00:21:30.881 "dma_device_id": "system", 00:21:30.881 "dma_device_type": 1 00:21:30.881 }, 00:21:30.881 { 00:21:30.881 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:30.881 "dma_device_type": 2 00:21:30.881 }, 00:21:30.881 { 00:21:30.881 "dma_device_id": "system", 00:21:30.881 "dma_device_type": 1 00:21:30.881 }, 00:21:30.881 { 00:21:30.881 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:30.881 "dma_device_type": 2 00:21:30.881 } 00:21:30.881 ], 00:21:30.881 "driver_specific": { 00:21:30.881 "raid": { 00:21:30.881 "uuid": "b4b71643-c1d2-4dd7-880c-cb7452c6aa75", 00:21:30.881 "strip_size_kb": 64, 00:21:30.881 "state": "online", 00:21:30.881 "raid_level": "raid0", 00:21:30.881 "superblock": false, 00:21:30.881 "num_base_bdevs": 3, 00:21:30.881 "num_base_bdevs_discovered": 3, 00:21:30.881 "num_base_bdevs_operational": 3, 00:21:30.881 "base_bdevs_list": [ 00:21:30.881 { 00:21:30.881 "name": "NewBaseBdev", 00:21:30.881 "uuid": "9910cc73-245b-4382-bf2e-f2bbe1c93a5a", 00:21:30.881 "is_configured": true, 00:21:30.881 "data_offset": 0, 00:21:30.881 "data_size": 65536 00:21:30.881 }, 00:21:30.881 { 00:21:30.881 "name": "BaseBdev2", 00:21:30.881 "uuid": "599cc638-1eb1-4faa-96bb-cd116bbc0119", 00:21:30.881 "is_configured": true, 00:21:30.881 "data_offset": 0, 00:21:30.881 "data_size": 65536 00:21:30.881 }, 00:21:30.881 { 00:21:30.881 "name": "BaseBdev3", 00:21:30.881 "uuid": "1de04ba4-0769-4569-8949-baa4f604a8b6", 00:21:30.881 "is_configured": true, 00:21:30.881 "data_offset": 0, 00:21:30.881 "data_size": 65536 00:21:30.881 } 00:21:30.881 ] 00:21:30.881 } 00:21:30.881 } 00:21:30.881 }' 00:21:30.881 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:21:31.140 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:21:31.140 BaseBdev2 00:21:31.140 BaseBdev3' 00:21:31.140 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:31.141 [2024-10-01 20:20:26.359113] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:21:31.141 [2024-10-01 20:20:26.359296] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:21:31.141 [2024-10-01 20:20:26.359511] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:21:31.141 [2024-10-01 20:20:26.359765] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:21:31.141 [2024-10-01 20:20:26.359800] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 64307 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 64307 ']' 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 64307 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:21:31.141 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 64307 00:21:31.399 killing process with pid 64307 00:21:31.399 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:21:31.399 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:21:31.399 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 64307' 00:21:31.399 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 64307 00:21:31.399 [2024-10-01 20:20:26.396964] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:21:31.399 20:20:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 64307 00:21:31.658 [2024-10-01 20:20:26.673454] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:21:33.559 00:21:33.559 real 0m12.874s 00:21:33.559 user 0m20.668s 00:21:33.559 sys 0m1.805s 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:21:33.559 ************************************ 00:21:33.559 END TEST raid_state_function_test 00:21:33.559 ************************************ 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:21:33.559 20:20:28 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 3 true 00:21:33.559 20:20:28 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:21:33.559 20:20:28 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:21:33.559 20:20:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:21:33.559 ************************************ 00:21:33.559 START TEST raid_state_function_test_sb 00:21:33.559 ************************************ 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 3 true 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:21:33.559 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:21:33.560 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:21:33.560 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:21:33.560 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:21:33.560 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:21:33.560 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:21:33.560 Process raid pid: 64956 00:21:33.560 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=64956 00:21:33.560 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 64956' 00:21:33.560 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:21:33.560 20:20:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 64956 00:21:33.560 20:20:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 64956 ']' 00:21:33.560 20:20:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:33.560 20:20:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:21:33.560 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:33.560 20:20:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:33.560 20:20:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:21:33.560 20:20:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:33.560 [2024-10-01 20:20:28.693340] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:21:33.560 [2024-10-01 20:20:28.693819] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:33.817 [2024-10-01 20:20:28.873031] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:34.075 [2024-10-01 20:20:29.159997] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:21:34.332 [2024-10-01 20:20:29.389197] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:34.332 [2024-10-01 20:20:29.389253] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:34.899 [2024-10-01 20:20:29.908465] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:21:34.899 [2024-10-01 20:20:29.908748] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:21:34.899 [2024-10-01 20:20:29.908933] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:21:34.899 [2024-10-01 20:20:29.909013] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:21:34.899 [2024-10-01 20:20:29.909265] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:21:34.899 [2024-10-01 20:20:29.909306] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:34.899 "name": "Existed_Raid", 00:21:34.899 "uuid": "4b2c92d8-88f8-4a70-8399-6da6fe42d423", 00:21:34.899 "strip_size_kb": 64, 00:21:34.899 "state": "configuring", 00:21:34.899 "raid_level": "raid0", 00:21:34.899 "superblock": true, 00:21:34.899 "num_base_bdevs": 3, 00:21:34.899 "num_base_bdevs_discovered": 0, 00:21:34.899 "num_base_bdevs_operational": 3, 00:21:34.899 "base_bdevs_list": [ 00:21:34.899 { 00:21:34.899 "name": "BaseBdev1", 00:21:34.899 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:34.899 "is_configured": false, 00:21:34.899 "data_offset": 0, 00:21:34.899 "data_size": 0 00:21:34.899 }, 00:21:34.899 { 00:21:34.899 "name": "BaseBdev2", 00:21:34.899 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:34.899 "is_configured": false, 00:21:34.899 "data_offset": 0, 00:21:34.899 "data_size": 0 00:21:34.899 }, 00:21:34.899 { 00:21:34.899 "name": "BaseBdev3", 00:21:34.899 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:34.899 "is_configured": false, 00:21:34.899 "data_offset": 0, 00:21:34.899 "data_size": 0 00:21:34.899 } 00:21:34.899 ] 00:21:34.899 }' 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:34.899 20:20:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:35.466 [2024-10-01 20:20:30.448479] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:21:35.466 [2024-10-01 20:20:30.448716] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:35.466 [2024-10-01 20:20:30.460500] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:21:35.466 [2024-10-01 20:20:30.460746] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:21:35.466 [2024-10-01 20:20:30.460892] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:21:35.466 [2024-10-01 20:20:30.460958] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:21:35.466 [2024-10-01 20:20:30.461087] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:21:35.466 [2024-10-01 20:20:30.461233] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:35.466 [2024-10-01 20:20:30.506367] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:21:35.466 BaseBdev1 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:35.466 [ 00:21:35.466 { 00:21:35.466 "name": "BaseBdev1", 00:21:35.466 "aliases": [ 00:21:35.466 "2f196dd5-0e89-420a-9da4-7fc39ef7b2ce" 00:21:35.466 ], 00:21:35.466 "product_name": "Malloc disk", 00:21:35.466 "block_size": 512, 00:21:35.466 "num_blocks": 65536, 00:21:35.466 "uuid": "2f196dd5-0e89-420a-9da4-7fc39ef7b2ce", 00:21:35.466 "assigned_rate_limits": { 00:21:35.466 "rw_ios_per_sec": 0, 00:21:35.466 "rw_mbytes_per_sec": 0, 00:21:35.466 "r_mbytes_per_sec": 0, 00:21:35.466 "w_mbytes_per_sec": 0 00:21:35.466 }, 00:21:35.466 "claimed": true, 00:21:35.466 "claim_type": "exclusive_write", 00:21:35.466 "zoned": false, 00:21:35.466 "supported_io_types": { 00:21:35.466 "read": true, 00:21:35.466 "write": true, 00:21:35.466 "unmap": true, 00:21:35.466 "flush": true, 00:21:35.466 "reset": true, 00:21:35.466 "nvme_admin": false, 00:21:35.466 "nvme_io": false, 00:21:35.466 "nvme_io_md": false, 00:21:35.466 "write_zeroes": true, 00:21:35.466 "zcopy": true, 00:21:35.466 "get_zone_info": false, 00:21:35.466 "zone_management": false, 00:21:35.466 "zone_append": false, 00:21:35.466 "compare": false, 00:21:35.466 "compare_and_write": false, 00:21:35.466 "abort": true, 00:21:35.466 "seek_hole": false, 00:21:35.466 "seek_data": false, 00:21:35.466 "copy": true, 00:21:35.466 "nvme_iov_md": false 00:21:35.466 }, 00:21:35.466 "memory_domains": [ 00:21:35.466 { 00:21:35.466 "dma_device_id": "system", 00:21:35.466 "dma_device_type": 1 00:21:35.466 }, 00:21:35.466 { 00:21:35.466 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:35.466 "dma_device_type": 2 00:21:35.466 } 00:21:35.466 ], 00:21:35.466 "driver_specific": {} 00:21:35.466 } 00:21:35.466 ] 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:35.466 "name": "Existed_Raid", 00:21:35.466 "uuid": "465c51f3-a669-4ee2-b23b-576073c8d96b", 00:21:35.466 "strip_size_kb": 64, 00:21:35.466 "state": "configuring", 00:21:35.466 "raid_level": "raid0", 00:21:35.466 "superblock": true, 00:21:35.466 "num_base_bdevs": 3, 00:21:35.466 "num_base_bdevs_discovered": 1, 00:21:35.466 "num_base_bdevs_operational": 3, 00:21:35.466 "base_bdevs_list": [ 00:21:35.466 { 00:21:35.466 "name": "BaseBdev1", 00:21:35.466 "uuid": "2f196dd5-0e89-420a-9da4-7fc39ef7b2ce", 00:21:35.466 "is_configured": true, 00:21:35.466 "data_offset": 2048, 00:21:35.466 "data_size": 63488 00:21:35.466 }, 00:21:35.466 { 00:21:35.466 "name": "BaseBdev2", 00:21:35.466 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:35.466 "is_configured": false, 00:21:35.466 "data_offset": 0, 00:21:35.466 "data_size": 0 00:21:35.466 }, 00:21:35.466 { 00:21:35.466 "name": "BaseBdev3", 00:21:35.466 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:35.466 "is_configured": false, 00:21:35.466 "data_offset": 0, 00:21:35.466 "data_size": 0 00:21:35.466 } 00:21:35.466 ] 00:21:35.466 }' 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:35.466 20:20:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:36.035 [2024-10-01 20:20:31.110653] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:21:36.035 [2024-10-01 20:20:31.110720] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:36.035 [2024-10-01 20:20:31.118694] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:21:36.035 [2024-10-01 20:20:31.121514] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:21:36.035 [2024-10-01 20:20:31.121754] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:21:36.035 [2024-10-01 20:20:31.121783] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:21:36.035 [2024-10-01 20:20:31.121803] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:36.035 "name": "Existed_Raid", 00:21:36.035 "uuid": "c602a7e5-8e38-4d63-8c5e-68955366cf1b", 00:21:36.035 "strip_size_kb": 64, 00:21:36.035 "state": "configuring", 00:21:36.035 "raid_level": "raid0", 00:21:36.035 "superblock": true, 00:21:36.035 "num_base_bdevs": 3, 00:21:36.035 "num_base_bdevs_discovered": 1, 00:21:36.035 "num_base_bdevs_operational": 3, 00:21:36.035 "base_bdevs_list": [ 00:21:36.035 { 00:21:36.035 "name": "BaseBdev1", 00:21:36.035 "uuid": "2f196dd5-0e89-420a-9da4-7fc39ef7b2ce", 00:21:36.035 "is_configured": true, 00:21:36.035 "data_offset": 2048, 00:21:36.035 "data_size": 63488 00:21:36.035 }, 00:21:36.035 { 00:21:36.035 "name": "BaseBdev2", 00:21:36.035 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:36.035 "is_configured": false, 00:21:36.035 "data_offset": 0, 00:21:36.035 "data_size": 0 00:21:36.035 }, 00:21:36.035 { 00:21:36.035 "name": "BaseBdev3", 00:21:36.035 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:36.035 "is_configured": false, 00:21:36.035 "data_offset": 0, 00:21:36.035 "data_size": 0 00:21:36.035 } 00:21:36.035 ] 00:21:36.035 }' 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:36.035 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:36.601 [2024-10-01 20:20:31.690696] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:21:36.601 BaseBdev2 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:36.601 [ 00:21:36.601 { 00:21:36.601 "name": "BaseBdev2", 00:21:36.601 "aliases": [ 00:21:36.601 "95aac73d-2554-4e96-a748-4997672e5bec" 00:21:36.601 ], 00:21:36.601 "product_name": "Malloc disk", 00:21:36.601 "block_size": 512, 00:21:36.601 "num_blocks": 65536, 00:21:36.601 "uuid": "95aac73d-2554-4e96-a748-4997672e5bec", 00:21:36.601 "assigned_rate_limits": { 00:21:36.601 "rw_ios_per_sec": 0, 00:21:36.601 "rw_mbytes_per_sec": 0, 00:21:36.601 "r_mbytes_per_sec": 0, 00:21:36.601 "w_mbytes_per_sec": 0 00:21:36.601 }, 00:21:36.601 "claimed": true, 00:21:36.601 "claim_type": "exclusive_write", 00:21:36.601 "zoned": false, 00:21:36.601 "supported_io_types": { 00:21:36.601 "read": true, 00:21:36.601 "write": true, 00:21:36.601 "unmap": true, 00:21:36.601 "flush": true, 00:21:36.601 "reset": true, 00:21:36.601 "nvme_admin": false, 00:21:36.601 "nvme_io": false, 00:21:36.601 "nvme_io_md": false, 00:21:36.601 "write_zeroes": true, 00:21:36.601 "zcopy": true, 00:21:36.601 "get_zone_info": false, 00:21:36.601 "zone_management": false, 00:21:36.601 "zone_append": false, 00:21:36.601 "compare": false, 00:21:36.601 "compare_and_write": false, 00:21:36.601 "abort": true, 00:21:36.601 "seek_hole": false, 00:21:36.601 "seek_data": false, 00:21:36.601 "copy": true, 00:21:36.601 "nvme_iov_md": false 00:21:36.601 }, 00:21:36.601 "memory_domains": [ 00:21:36.601 { 00:21:36.601 "dma_device_id": "system", 00:21:36.601 "dma_device_type": 1 00:21:36.601 }, 00:21:36.601 { 00:21:36.601 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:36.601 "dma_device_type": 2 00:21:36.601 } 00:21:36.601 ], 00:21:36.601 "driver_specific": {} 00:21:36.601 } 00:21:36.601 ] 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:36.601 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:36.602 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:36.602 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:36.602 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:36.602 "name": "Existed_Raid", 00:21:36.602 "uuid": "c602a7e5-8e38-4d63-8c5e-68955366cf1b", 00:21:36.602 "strip_size_kb": 64, 00:21:36.602 "state": "configuring", 00:21:36.602 "raid_level": "raid0", 00:21:36.602 "superblock": true, 00:21:36.602 "num_base_bdevs": 3, 00:21:36.602 "num_base_bdevs_discovered": 2, 00:21:36.602 "num_base_bdevs_operational": 3, 00:21:36.602 "base_bdevs_list": [ 00:21:36.602 { 00:21:36.602 "name": "BaseBdev1", 00:21:36.602 "uuid": "2f196dd5-0e89-420a-9da4-7fc39ef7b2ce", 00:21:36.602 "is_configured": true, 00:21:36.602 "data_offset": 2048, 00:21:36.602 "data_size": 63488 00:21:36.602 }, 00:21:36.602 { 00:21:36.602 "name": "BaseBdev2", 00:21:36.602 "uuid": "95aac73d-2554-4e96-a748-4997672e5bec", 00:21:36.602 "is_configured": true, 00:21:36.602 "data_offset": 2048, 00:21:36.602 "data_size": 63488 00:21:36.602 }, 00:21:36.602 { 00:21:36.602 "name": "BaseBdev3", 00:21:36.602 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:36.602 "is_configured": false, 00:21:36.602 "data_offset": 0, 00:21:36.602 "data_size": 0 00:21:36.602 } 00:21:36.602 ] 00:21:36.602 }' 00:21:36.602 20:20:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:36.602 20:20:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:37.166 [2024-10-01 20:20:32.272466] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:21:37.166 [2024-10-01 20:20:32.272846] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:21:37.166 [2024-10-01 20:20:32.272903] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:21:37.166 BaseBdev3 00:21:37.166 [2024-10-01 20:20:32.273301] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:21:37.166 [2024-10-01 20:20:32.273506] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:21:37.166 [2024-10-01 20:20:32.273523] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:21:37.166 [2024-10-01 20:20:32.273746] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:37.166 [ 00:21:37.166 { 00:21:37.166 "name": "BaseBdev3", 00:21:37.166 "aliases": [ 00:21:37.166 "ee419a57-d584-460d-92f4-c573f34ce270" 00:21:37.166 ], 00:21:37.166 "product_name": "Malloc disk", 00:21:37.166 "block_size": 512, 00:21:37.166 "num_blocks": 65536, 00:21:37.166 "uuid": "ee419a57-d584-460d-92f4-c573f34ce270", 00:21:37.166 "assigned_rate_limits": { 00:21:37.166 "rw_ios_per_sec": 0, 00:21:37.166 "rw_mbytes_per_sec": 0, 00:21:37.166 "r_mbytes_per_sec": 0, 00:21:37.166 "w_mbytes_per_sec": 0 00:21:37.166 }, 00:21:37.166 "claimed": true, 00:21:37.166 "claim_type": "exclusive_write", 00:21:37.166 "zoned": false, 00:21:37.166 "supported_io_types": { 00:21:37.166 "read": true, 00:21:37.166 "write": true, 00:21:37.166 "unmap": true, 00:21:37.166 "flush": true, 00:21:37.166 "reset": true, 00:21:37.166 "nvme_admin": false, 00:21:37.166 "nvme_io": false, 00:21:37.166 "nvme_io_md": false, 00:21:37.166 "write_zeroes": true, 00:21:37.166 "zcopy": true, 00:21:37.166 "get_zone_info": false, 00:21:37.166 "zone_management": false, 00:21:37.166 "zone_append": false, 00:21:37.166 "compare": false, 00:21:37.166 "compare_and_write": false, 00:21:37.166 "abort": true, 00:21:37.166 "seek_hole": false, 00:21:37.166 "seek_data": false, 00:21:37.166 "copy": true, 00:21:37.166 "nvme_iov_md": false 00:21:37.166 }, 00:21:37.166 "memory_domains": [ 00:21:37.166 { 00:21:37.166 "dma_device_id": "system", 00:21:37.166 "dma_device_type": 1 00:21:37.166 }, 00:21:37.166 { 00:21:37.166 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:37.166 "dma_device_type": 2 00:21:37.166 } 00:21:37.166 ], 00:21:37.166 "driver_specific": {} 00:21:37.166 } 00:21:37.166 ] 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:37.166 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:37.167 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:37.167 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:37.167 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:37.167 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:37.167 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:37.167 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:37.167 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:37.167 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.167 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:37.167 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.167 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:37.167 "name": "Existed_Raid", 00:21:37.167 "uuid": "c602a7e5-8e38-4d63-8c5e-68955366cf1b", 00:21:37.167 "strip_size_kb": 64, 00:21:37.167 "state": "online", 00:21:37.167 "raid_level": "raid0", 00:21:37.167 "superblock": true, 00:21:37.167 "num_base_bdevs": 3, 00:21:37.167 "num_base_bdevs_discovered": 3, 00:21:37.167 "num_base_bdevs_operational": 3, 00:21:37.167 "base_bdevs_list": [ 00:21:37.167 { 00:21:37.167 "name": "BaseBdev1", 00:21:37.167 "uuid": "2f196dd5-0e89-420a-9da4-7fc39ef7b2ce", 00:21:37.167 "is_configured": true, 00:21:37.167 "data_offset": 2048, 00:21:37.167 "data_size": 63488 00:21:37.167 }, 00:21:37.167 { 00:21:37.167 "name": "BaseBdev2", 00:21:37.167 "uuid": "95aac73d-2554-4e96-a748-4997672e5bec", 00:21:37.167 "is_configured": true, 00:21:37.167 "data_offset": 2048, 00:21:37.167 "data_size": 63488 00:21:37.167 }, 00:21:37.167 { 00:21:37.167 "name": "BaseBdev3", 00:21:37.167 "uuid": "ee419a57-d584-460d-92f4-c573f34ce270", 00:21:37.167 "is_configured": true, 00:21:37.167 "data_offset": 2048, 00:21:37.167 "data_size": 63488 00:21:37.167 } 00:21:37.167 ] 00:21:37.167 }' 00:21:37.167 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:37.167 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:37.733 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:21:37.733 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:21:37.733 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:21:37.733 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:21:37.733 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:21:37.733 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:21:37.733 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:21:37.733 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:21:37.733 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.733 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:37.733 [2024-10-01 20:20:32.849138] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:21:37.733 20:20:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.733 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:21:37.733 "name": "Existed_Raid", 00:21:37.733 "aliases": [ 00:21:37.733 "c602a7e5-8e38-4d63-8c5e-68955366cf1b" 00:21:37.733 ], 00:21:37.733 "product_name": "Raid Volume", 00:21:37.733 "block_size": 512, 00:21:37.733 "num_blocks": 190464, 00:21:37.733 "uuid": "c602a7e5-8e38-4d63-8c5e-68955366cf1b", 00:21:37.733 "assigned_rate_limits": { 00:21:37.733 "rw_ios_per_sec": 0, 00:21:37.733 "rw_mbytes_per_sec": 0, 00:21:37.733 "r_mbytes_per_sec": 0, 00:21:37.733 "w_mbytes_per_sec": 0 00:21:37.733 }, 00:21:37.733 "claimed": false, 00:21:37.733 "zoned": false, 00:21:37.733 "supported_io_types": { 00:21:37.733 "read": true, 00:21:37.733 "write": true, 00:21:37.733 "unmap": true, 00:21:37.733 "flush": true, 00:21:37.733 "reset": true, 00:21:37.733 "nvme_admin": false, 00:21:37.733 "nvme_io": false, 00:21:37.733 "nvme_io_md": false, 00:21:37.733 "write_zeroes": true, 00:21:37.733 "zcopy": false, 00:21:37.733 "get_zone_info": false, 00:21:37.733 "zone_management": false, 00:21:37.733 "zone_append": false, 00:21:37.733 "compare": false, 00:21:37.733 "compare_and_write": false, 00:21:37.733 "abort": false, 00:21:37.733 "seek_hole": false, 00:21:37.733 "seek_data": false, 00:21:37.733 "copy": false, 00:21:37.733 "nvme_iov_md": false 00:21:37.733 }, 00:21:37.733 "memory_domains": [ 00:21:37.733 { 00:21:37.733 "dma_device_id": "system", 00:21:37.733 "dma_device_type": 1 00:21:37.733 }, 00:21:37.733 { 00:21:37.733 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:37.733 "dma_device_type": 2 00:21:37.733 }, 00:21:37.733 { 00:21:37.733 "dma_device_id": "system", 00:21:37.733 "dma_device_type": 1 00:21:37.733 }, 00:21:37.733 { 00:21:37.733 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:37.733 "dma_device_type": 2 00:21:37.733 }, 00:21:37.733 { 00:21:37.733 "dma_device_id": "system", 00:21:37.733 "dma_device_type": 1 00:21:37.733 }, 00:21:37.733 { 00:21:37.733 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:37.733 "dma_device_type": 2 00:21:37.733 } 00:21:37.733 ], 00:21:37.733 "driver_specific": { 00:21:37.733 "raid": { 00:21:37.733 "uuid": "c602a7e5-8e38-4d63-8c5e-68955366cf1b", 00:21:37.733 "strip_size_kb": 64, 00:21:37.733 "state": "online", 00:21:37.733 "raid_level": "raid0", 00:21:37.733 "superblock": true, 00:21:37.733 "num_base_bdevs": 3, 00:21:37.733 "num_base_bdevs_discovered": 3, 00:21:37.733 "num_base_bdevs_operational": 3, 00:21:37.733 "base_bdevs_list": [ 00:21:37.733 { 00:21:37.733 "name": "BaseBdev1", 00:21:37.733 "uuid": "2f196dd5-0e89-420a-9da4-7fc39ef7b2ce", 00:21:37.733 "is_configured": true, 00:21:37.733 "data_offset": 2048, 00:21:37.733 "data_size": 63488 00:21:37.733 }, 00:21:37.733 { 00:21:37.733 "name": "BaseBdev2", 00:21:37.733 "uuid": "95aac73d-2554-4e96-a748-4997672e5bec", 00:21:37.733 "is_configured": true, 00:21:37.733 "data_offset": 2048, 00:21:37.733 "data_size": 63488 00:21:37.733 }, 00:21:37.733 { 00:21:37.733 "name": "BaseBdev3", 00:21:37.733 "uuid": "ee419a57-d584-460d-92f4-c573f34ce270", 00:21:37.733 "is_configured": true, 00:21:37.733 "data_offset": 2048, 00:21:37.733 "data_size": 63488 00:21:37.733 } 00:21:37.733 ] 00:21:37.733 } 00:21:37.733 } 00:21:37.733 }' 00:21:37.733 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:21:37.733 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:21:37.733 BaseBdev2 00:21:37.733 BaseBdev3' 00:21:37.733 20:20:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:21:37.992 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.993 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:37.993 [2024-10-01 20:20:33.168886] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:21:37.993 [2024-10-01 20:20:33.169068] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:21:37.993 [2024-10-01 20:20:33.169266] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:38.251 "name": "Existed_Raid", 00:21:38.251 "uuid": "c602a7e5-8e38-4d63-8c5e-68955366cf1b", 00:21:38.251 "strip_size_kb": 64, 00:21:38.251 "state": "offline", 00:21:38.251 "raid_level": "raid0", 00:21:38.251 "superblock": true, 00:21:38.251 "num_base_bdevs": 3, 00:21:38.251 "num_base_bdevs_discovered": 2, 00:21:38.251 "num_base_bdevs_operational": 2, 00:21:38.251 "base_bdevs_list": [ 00:21:38.251 { 00:21:38.251 "name": null, 00:21:38.251 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:38.251 "is_configured": false, 00:21:38.251 "data_offset": 0, 00:21:38.251 "data_size": 63488 00:21:38.251 }, 00:21:38.251 { 00:21:38.251 "name": "BaseBdev2", 00:21:38.251 "uuid": "95aac73d-2554-4e96-a748-4997672e5bec", 00:21:38.251 "is_configured": true, 00:21:38.251 "data_offset": 2048, 00:21:38.251 "data_size": 63488 00:21:38.251 }, 00:21:38.251 { 00:21:38.251 "name": "BaseBdev3", 00:21:38.251 "uuid": "ee419a57-d584-460d-92f4-c573f34ce270", 00:21:38.251 "is_configured": true, 00:21:38.251 "data_offset": 2048, 00:21:38.251 "data_size": 63488 00:21:38.251 } 00:21:38.251 ] 00:21:38.251 }' 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:38.251 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:38.818 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:21:38.818 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:21:38.818 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:38.818 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:38.819 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:38.819 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:21:38.819 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:38.819 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:21:38.819 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:21:38.819 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:21:38.819 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:38.819 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:38.819 [2024-10-01 20:20:33.865956] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:21:38.819 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:38.819 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:21:38.819 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:21:38.819 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:38.819 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:38.819 20:20:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:21:38.819 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:38.819 20:20:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:38.819 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:21:38.819 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:21:38.819 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:21:38.819 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:38.819 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:38.819 [2024-10-01 20:20:34.016220] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:21:38.819 [2024-10-01 20:20:34.016545] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:39.078 BaseBdev2 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:39.078 [ 00:21:39.078 { 00:21:39.078 "name": "BaseBdev2", 00:21:39.078 "aliases": [ 00:21:39.078 "3d52c515-bac5-4b57-b22d-4fb4754b399f" 00:21:39.078 ], 00:21:39.078 "product_name": "Malloc disk", 00:21:39.078 "block_size": 512, 00:21:39.078 "num_blocks": 65536, 00:21:39.078 "uuid": "3d52c515-bac5-4b57-b22d-4fb4754b399f", 00:21:39.078 "assigned_rate_limits": { 00:21:39.078 "rw_ios_per_sec": 0, 00:21:39.078 "rw_mbytes_per_sec": 0, 00:21:39.078 "r_mbytes_per_sec": 0, 00:21:39.078 "w_mbytes_per_sec": 0 00:21:39.078 }, 00:21:39.078 "claimed": false, 00:21:39.078 "zoned": false, 00:21:39.078 "supported_io_types": { 00:21:39.078 "read": true, 00:21:39.078 "write": true, 00:21:39.078 "unmap": true, 00:21:39.078 "flush": true, 00:21:39.078 "reset": true, 00:21:39.078 "nvme_admin": false, 00:21:39.078 "nvme_io": false, 00:21:39.078 "nvme_io_md": false, 00:21:39.078 "write_zeroes": true, 00:21:39.078 "zcopy": true, 00:21:39.078 "get_zone_info": false, 00:21:39.078 "zone_management": false, 00:21:39.078 "zone_append": false, 00:21:39.078 "compare": false, 00:21:39.078 "compare_and_write": false, 00:21:39.078 "abort": true, 00:21:39.078 "seek_hole": false, 00:21:39.078 "seek_data": false, 00:21:39.078 "copy": true, 00:21:39.078 "nvme_iov_md": false 00:21:39.078 }, 00:21:39.078 "memory_domains": [ 00:21:39.078 { 00:21:39.078 "dma_device_id": "system", 00:21:39.078 "dma_device_type": 1 00:21:39.078 }, 00:21:39.078 { 00:21:39.078 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:39.078 "dma_device_type": 2 00:21:39.078 } 00:21:39.078 ], 00:21:39.078 "driver_specific": {} 00:21:39.078 } 00:21:39.078 ] 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:39.078 BaseBdev3 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:39.078 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:39.078 [ 00:21:39.078 { 00:21:39.078 "name": "BaseBdev3", 00:21:39.078 "aliases": [ 00:21:39.078 "1a372f20-1132-4d71-91f3-2611e13f28f7" 00:21:39.078 ], 00:21:39.078 "product_name": "Malloc disk", 00:21:39.078 "block_size": 512, 00:21:39.078 "num_blocks": 65536, 00:21:39.078 "uuid": "1a372f20-1132-4d71-91f3-2611e13f28f7", 00:21:39.078 "assigned_rate_limits": { 00:21:39.078 "rw_ios_per_sec": 0, 00:21:39.078 "rw_mbytes_per_sec": 0, 00:21:39.078 "r_mbytes_per_sec": 0, 00:21:39.078 "w_mbytes_per_sec": 0 00:21:39.078 }, 00:21:39.078 "claimed": false, 00:21:39.078 "zoned": false, 00:21:39.078 "supported_io_types": { 00:21:39.078 "read": true, 00:21:39.078 "write": true, 00:21:39.078 "unmap": true, 00:21:39.078 "flush": true, 00:21:39.078 "reset": true, 00:21:39.078 "nvme_admin": false, 00:21:39.078 "nvme_io": false, 00:21:39.078 "nvme_io_md": false, 00:21:39.078 "write_zeroes": true, 00:21:39.078 "zcopy": true, 00:21:39.078 "get_zone_info": false, 00:21:39.078 "zone_management": false, 00:21:39.078 "zone_append": false, 00:21:39.078 "compare": false, 00:21:39.078 "compare_and_write": false, 00:21:39.078 "abort": true, 00:21:39.078 "seek_hole": false, 00:21:39.078 "seek_data": false, 00:21:39.078 "copy": true, 00:21:39.078 "nvme_iov_md": false 00:21:39.078 }, 00:21:39.078 "memory_domains": [ 00:21:39.078 { 00:21:39.078 "dma_device_id": "system", 00:21:39.078 "dma_device_type": 1 00:21:39.078 }, 00:21:39.078 { 00:21:39.078 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:39.078 "dma_device_type": 2 00:21:39.078 } 00:21:39.078 ], 00:21:39.078 "driver_specific": {} 00:21:39.078 } 00:21:39.079 ] 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:39.079 [2024-10-01 20:20:34.309901] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:21:39.079 [2024-10-01 20:20:34.310098] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:21:39.079 [2024-10-01 20:20:34.310241] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:21:39.079 [2024-10-01 20:20:34.312914] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:39.079 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:39.337 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:39.337 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:39.337 "name": "Existed_Raid", 00:21:39.337 "uuid": "6d9ded77-d46e-4400-9cd9-609d47bdc4e6", 00:21:39.337 "strip_size_kb": 64, 00:21:39.337 "state": "configuring", 00:21:39.337 "raid_level": "raid0", 00:21:39.337 "superblock": true, 00:21:39.337 "num_base_bdevs": 3, 00:21:39.337 "num_base_bdevs_discovered": 2, 00:21:39.337 "num_base_bdevs_operational": 3, 00:21:39.337 "base_bdevs_list": [ 00:21:39.337 { 00:21:39.337 "name": "BaseBdev1", 00:21:39.337 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:39.337 "is_configured": false, 00:21:39.337 "data_offset": 0, 00:21:39.337 "data_size": 0 00:21:39.337 }, 00:21:39.337 { 00:21:39.337 "name": "BaseBdev2", 00:21:39.337 "uuid": "3d52c515-bac5-4b57-b22d-4fb4754b399f", 00:21:39.337 "is_configured": true, 00:21:39.337 "data_offset": 2048, 00:21:39.337 "data_size": 63488 00:21:39.337 }, 00:21:39.337 { 00:21:39.337 "name": "BaseBdev3", 00:21:39.337 "uuid": "1a372f20-1132-4d71-91f3-2611e13f28f7", 00:21:39.337 "is_configured": true, 00:21:39.337 "data_offset": 2048, 00:21:39.337 "data_size": 63488 00:21:39.337 } 00:21:39.337 ] 00:21:39.337 }' 00:21:39.337 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:39.337 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:39.596 [2024-10-01 20:20:34.833982] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:39.596 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:39.855 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:39.855 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:39.855 "name": "Existed_Raid", 00:21:39.855 "uuid": "6d9ded77-d46e-4400-9cd9-609d47bdc4e6", 00:21:39.855 "strip_size_kb": 64, 00:21:39.855 "state": "configuring", 00:21:39.855 "raid_level": "raid0", 00:21:39.855 "superblock": true, 00:21:39.855 "num_base_bdevs": 3, 00:21:39.855 "num_base_bdevs_discovered": 1, 00:21:39.855 "num_base_bdevs_operational": 3, 00:21:39.855 "base_bdevs_list": [ 00:21:39.855 { 00:21:39.855 "name": "BaseBdev1", 00:21:39.855 "uuid": "00000000-0000-0000-0000-000000000000", 00:21:39.855 "is_configured": false, 00:21:39.855 "data_offset": 0, 00:21:39.855 "data_size": 0 00:21:39.855 }, 00:21:39.855 { 00:21:39.855 "name": null, 00:21:39.855 "uuid": "3d52c515-bac5-4b57-b22d-4fb4754b399f", 00:21:39.855 "is_configured": false, 00:21:39.855 "data_offset": 0, 00:21:39.855 "data_size": 63488 00:21:39.855 }, 00:21:39.855 { 00:21:39.855 "name": "BaseBdev3", 00:21:39.855 "uuid": "1a372f20-1132-4d71-91f3-2611e13f28f7", 00:21:39.855 "is_configured": true, 00:21:39.855 "data_offset": 2048, 00:21:39.855 "data_size": 63488 00:21:39.855 } 00:21:39.855 ] 00:21:39.855 }' 00:21:39.855 20:20:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:39.855 20:20:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:40.113 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:40.113 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:40.113 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:21:40.113 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:40.113 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:40.372 BaseBdev1 00:21:40.372 [2024-10-01 20:20:35.429023] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:40.372 [ 00:21:40.372 { 00:21:40.372 "name": "BaseBdev1", 00:21:40.372 "aliases": [ 00:21:40.372 "0f88b71f-399d-4bf6-a08b-abc8dfd92235" 00:21:40.372 ], 00:21:40.372 "product_name": "Malloc disk", 00:21:40.372 "block_size": 512, 00:21:40.372 "num_blocks": 65536, 00:21:40.372 "uuid": "0f88b71f-399d-4bf6-a08b-abc8dfd92235", 00:21:40.372 "assigned_rate_limits": { 00:21:40.372 "rw_ios_per_sec": 0, 00:21:40.372 "rw_mbytes_per_sec": 0, 00:21:40.372 "r_mbytes_per_sec": 0, 00:21:40.372 "w_mbytes_per_sec": 0 00:21:40.372 }, 00:21:40.372 "claimed": true, 00:21:40.372 "claim_type": "exclusive_write", 00:21:40.372 "zoned": false, 00:21:40.372 "supported_io_types": { 00:21:40.372 "read": true, 00:21:40.372 "write": true, 00:21:40.372 "unmap": true, 00:21:40.372 "flush": true, 00:21:40.372 "reset": true, 00:21:40.372 "nvme_admin": false, 00:21:40.372 "nvme_io": false, 00:21:40.372 "nvme_io_md": false, 00:21:40.372 "write_zeroes": true, 00:21:40.372 "zcopy": true, 00:21:40.372 "get_zone_info": false, 00:21:40.372 "zone_management": false, 00:21:40.372 "zone_append": false, 00:21:40.372 "compare": false, 00:21:40.372 "compare_and_write": false, 00:21:40.372 "abort": true, 00:21:40.372 "seek_hole": false, 00:21:40.372 "seek_data": false, 00:21:40.372 "copy": true, 00:21:40.372 "nvme_iov_md": false 00:21:40.372 }, 00:21:40.372 "memory_domains": [ 00:21:40.372 { 00:21:40.372 "dma_device_id": "system", 00:21:40.372 "dma_device_type": 1 00:21:40.372 }, 00:21:40.372 { 00:21:40.372 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:40.372 "dma_device_type": 2 00:21:40.372 } 00:21:40.372 ], 00:21:40.372 "driver_specific": {} 00:21:40.372 } 00:21:40.372 ] 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:40.372 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:40.373 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:40.373 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:40.373 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:40.373 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:40.373 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:40.373 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:40.373 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:40.373 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:40.373 "name": "Existed_Raid", 00:21:40.373 "uuid": "6d9ded77-d46e-4400-9cd9-609d47bdc4e6", 00:21:40.373 "strip_size_kb": 64, 00:21:40.373 "state": "configuring", 00:21:40.373 "raid_level": "raid0", 00:21:40.373 "superblock": true, 00:21:40.373 "num_base_bdevs": 3, 00:21:40.373 "num_base_bdevs_discovered": 2, 00:21:40.373 "num_base_bdevs_operational": 3, 00:21:40.373 "base_bdevs_list": [ 00:21:40.373 { 00:21:40.373 "name": "BaseBdev1", 00:21:40.373 "uuid": "0f88b71f-399d-4bf6-a08b-abc8dfd92235", 00:21:40.373 "is_configured": true, 00:21:40.373 "data_offset": 2048, 00:21:40.373 "data_size": 63488 00:21:40.373 }, 00:21:40.373 { 00:21:40.373 "name": null, 00:21:40.373 "uuid": "3d52c515-bac5-4b57-b22d-4fb4754b399f", 00:21:40.373 "is_configured": false, 00:21:40.373 "data_offset": 0, 00:21:40.373 "data_size": 63488 00:21:40.373 }, 00:21:40.373 { 00:21:40.373 "name": "BaseBdev3", 00:21:40.373 "uuid": "1a372f20-1132-4d71-91f3-2611e13f28f7", 00:21:40.373 "is_configured": true, 00:21:40.373 "data_offset": 2048, 00:21:40.373 "data_size": 63488 00:21:40.373 } 00:21:40.373 ] 00:21:40.373 }' 00:21:40.373 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:40.373 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:40.940 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:40.940 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:40.940 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:40.940 20:20:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:21:40.940 20:20:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:40.940 [2024-10-01 20:20:36.037352] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:40.940 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:40.941 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:40.941 "name": "Existed_Raid", 00:21:40.941 "uuid": "6d9ded77-d46e-4400-9cd9-609d47bdc4e6", 00:21:40.941 "strip_size_kb": 64, 00:21:40.941 "state": "configuring", 00:21:40.941 "raid_level": "raid0", 00:21:40.941 "superblock": true, 00:21:40.941 "num_base_bdevs": 3, 00:21:40.941 "num_base_bdevs_discovered": 1, 00:21:40.941 "num_base_bdevs_operational": 3, 00:21:40.941 "base_bdevs_list": [ 00:21:40.941 { 00:21:40.941 "name": "BaseBdev1", 00:21:40.941 "uuid": "0f88b71f-399d-4bf6-a08b-abc8dfd92235", 00:21:40.941 "is_configured": true, 00:21:40.941 "data_offset": 2048, 00:21:40.941 "data_size": 63488 00:21:40.941 }, 00:21:40.941 { 00:21:40.941 "name": null, 00:21:40.941 "uuid": "3d52c515-bac5-4b57-b22d-4fb4754b399f", 00:21:40.941 "is_configured": false, 00:21:40.941 "data_offset": 0, 00:21:40.941 "data_size": 63488 00:21:40.941 }, 00:21:40.941 { 00:21:40.941 "name": null, 00:21:40.941 "uuid": "1a372f20-1132-4d71-91f3-2611e13f28f7", 00:21:40.941 "is_configured": false, 00:21:40.941 "data_offset": 0, 00:21:40.941 "data_size": 63488 00:21:40.941 } 00:21:40.941 ] 00:21:40.941 }' 00:21:40.941 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:40.941 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:41.508 [2024-10-01 20:20:36.609474] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:41.508 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:41.508 "name": "Existed_Raid", 00:21:41.508 "uuid": "6d9ded77-d46e-4400-9cd9-609d47bdc4e6", 00:21:41.508 "strip_size_kb": 64, 00:21:41.508 "state": "configuring", 00:21:41.508 "raid_level": "raid0", 00:21:41.508 "superblock": true, 00:21:41.508 "num_base_bdevs": 3, 00:21:41.508 "num_base_bdevs_discovered": 2, 00:21:41.508 "num_base_bdevs_operational": 3, 00:21:41.508 "base_bdevs_list": [ 00:21:41.508 { 00:21:41.508 "name": "BaseBdev1", 00:21:41.508 "uuid": "0f88b71f-399d-4bf6-a08b-abc8dfd92235", 00:21:41.508 "is_configured": true, 00:21:41.508 "data_offset": 2048, 00:21:41.508 "data_size": 63488 00:21:41.508 }, 00:21:41.508 { 00:21:41.508 "name": null, 00:21:41.508 "uuid": "3d52c515-bac5-4b57-b22d-4fb4754b399f", 00:21:41.508 "is_configured": false, 00:21:41.508 "data_offset": 0, 00:21:41.508 "data_size": 63488 00:21:41.508 }, 00:21:41.508 { 00:21:41.508 "name": "BaseBdev3", 00:21:41.508 "uuid": "1a372f20-1132-4d71-91f3-2611e13f28f7", 00:21:41.508 "is_configured": true, 00:21:41.508 "data_offset": 2048, 00:21:41.508 "data_size": 63488 00:21:41.508 } 00:21:41.508 ] 00:21:41.508 }' 00:21:41.509 20:20:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:41.509 20:20:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:42.099 [2024-10-01 20:20:37.177809] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:42.099 "name": "Existed_Raid", 00:21:42.099 "uuid": "6d9ded77-d46e-4400-9cd9-609d47bdc4e6", 00:21:42.099 "strip_size_kb": 64, 00:21:42.099 "state": "configuring", 00:21:42.099 "raid_level": "raid0", 00:21:42.099 "superblock": true, 00:21:42.099 "num_base_bdevs": 3, 00:21:42.099 "num_base_bdevs_discovered": 1, 00:21:42.099 "num_base_bdevs_operational": 3, 00:21:42.099 "base_bdevs_list": [ 00:21:42.099 { 00:21:42.099 "name": null, 00:21:42.099 "uuid": "0f88b71f-399d-4bf6-a08b-abc8dfd92235", 00:21:42.099 "is_configured": false, 00:21:42.099 "data_offset": 0, 00:21:42.099 "data_size": 63488 00:21:42.099 }, 00:21:42.099 { 00:21:42.099 "name": null, 00:21:42.099 "uuid": "3d52c515-bac5-4b57-b22d-4fb4754b399f", 00:21:42.099 "is_configured": false, 00:21:42.099 "data_offset": 0, 00:21:42.099 "data_size": 63488 00:21:42.099 }, 00:21:42.099 { 00:21:42.099 "name": "BaseBdev3", 00:21:42.099 "uuid": "1a372f20-1132-4d71-91f3-2611e13f28f7", 00:21:42.099 "is_configured": true, 00:21:42.099 "data_offset": 2048, 00:21:42.099 "data_size": 63488 00:21:42.099 } 00:21:42.099 ] 00:21:42.099 }' 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:42.099 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:42.666 [2024-10-01 20:20:37.864591] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:42.666 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:42.925 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:42.925 "name": "Existed_Raid", 00:21:42.925 "uuid": "6d9ded77-d46e-4400-9cd9-609d47bdc4e6", 00:21:42.925 "strip_size_kb": 64, 00:21:42.925 "state": "configuring", 00:21:42.925 "raid_level": "raid0", 00:21:42.925 "superblock": true, 00:21:42.925 "num_base_bdevs": 3, 00:21:42.925 "num_base_bdevs_discovered": 2, 00:21:42.925 "num_base_bdevs_operational": 3, 00:21:42.925 "base_bdevs_list": [ 00:21:42.925 { 00:21:42.925 "name": null, 00:21:42.925 "uuid": "0f88b71f-399d-4bf6-a08b-abc8dfd92235", 00:21:42.925 "is_configured": false, 00:21:42.925 "data_offset": 0, 00:21:42.925 "data_size": 63488 00:21:42.925 }, 00:21:42.925 { 00:21:42.925 "name": "BaseBdev2", 00:21:42.925 "uuid": "3d52c515-bac5-4b57-b22d-4fb4754b399f", 00:21:42.925 "is_configured": true, 00:21:42.925 "data_offset": 2048, 00:21:42.925 "data_size": 63488 00:21:42.925 }, 00:21:42.925 { 00:21:42.925 "name": "BaseBdev3", 00:21:42.925 "uuid": "1a372f20-1132-4d71-91f3-2611e13f28f7", 00:21:42.925 "is_configured": true, 00:21:42.925 "data_offset": 2048, 00:21:42.925 "data_size": 63488 00:21:42.925 } 00:21:42.925 ] 00:21:42.925 }' 00:21:42.925 20:20:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:42.925 20:20:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:43.185 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:43.185 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:21:43.185 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:43.185 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:43.185 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 0f88b71f-399d-4bf6-a08b-abc8dfd92235 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:43.444 [2024-10-01 20:20:38.549347] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:21:43.444 [2024-10-01 20:20:38.549669] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:21:43.444 [2024-10-01 20:20:38.549694] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:21:43.444 NewBaseBdev 00:21:43.444 [2024-10-01 20:20:38.550067] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:21:43.444 [2024-10-01 20:20:38.550249] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:21:43.444 [2024-10-01 20:20:38.550265] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:21:43.444 [2024-10-01 20:20:38.550435] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:43.444 [ 00:21:43.444 { 00:21:43.444 "name": "NewBaseBdev", 00:21:43.444 "aliases": [ 00:21:43.444 "0f88b71f-399d-4bf6-a08b-abc8dfd92235" 00:21:43.444 ], 00:21:43.444 "product_name": "Malloc disk", 00:21:43.444 "block_size": 512, 00:21:43.444 "num_blocks": 65536, 00:21:43.444 "uuid": "0f88b71f-399d-4bf6-a08b-abc8dfd92235", 00:21:43.444 "assigned_rate_limits": { 00:21:43.444 "rw_ios_per_sec": 0, 00:21:43.444 "rw_mbytes_per_sec": 0, 00:21:43.444 "r_mbytes_per_sec": 0, 00:21:43.444 "w_mbytes_per_sec": 0 00:21:43.444 }, 00:21:43.444 "claimed": true, 00:21:43.444 "claim_type": "exclusive_write", 00:21:43.444 "zoned": false, 00:21:43.444 "supported_io_types": { 00:21:43.444 "read": true, 00:21:43.444 "write": true, 00:21:43.444 "unmap": true, 00:21:43.444 "flush": true, 00:21:43.444 "reset": true, 00:21:43.444 "nvme_admin": false, 00:21:43.444 "nvme_io": false, 00:21:43.444 "nvme_io_md": false, 00:21:43.444 "write_zeroes": true, 00:21:43.444 "zcopy": true, 00:21:43.444 "get_zone_info": false, 00:21:43.444 "zone_management": false, 00:21:43.444 "zone_append": false, 00:21:43.444 "compare": false, 00:21:43.444 "compare_and_write": false, 00:21:43.444 "abort": true, 00:21:43.444 "seek_hole": false, 00:21:43.444 "seek_data": false, 00:21:43.444 "copy": true, 00:21:43.444 "nvme_iov_md": false 00:21:43.444 }, 00:21:43.444 "memory_domains": [ 00:21:43.444 { 00:21:43.444 "dma_device_id": "system", 00:21:43.444 "dma_device_type": 1 00:21:43.444 }, 00:21:43.444 { 00:21:43.444 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:43.444 "dma_device_type": 2 00:21:43.444 } 00:21:43.444 ], 00:21:43.444 "driver_specific": {} 00:21:43.444 } 00:21:43.444 ] 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:43.444 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:43.445 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:43.445 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:21:43.445 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:43.445 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:43.445 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:43.445 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:43.445 "name": "Existed_Raid", 00:21:43.445 "uuid": "6d9ded77-d46e-4400-9cd9-609d47bdc4e6", 00:21:43.445 "strip_size_kb": 64, 00:21:43.445 "state": "online", 00:21:43.445 "raid_level": "raid0", 00:21:43.445 "superblock": true, 00:21:43.445 "num_base_bdevs": 3, 00:21:43.445 "num_base_bdevs_discovered": 3, 00:21:43.445 "num_base_bdevs_operational": 3, 00:21:43.445 "base_bdevs_list": [ 00:21:43.445 { 00:21:43.445 "name": "NewBaseBdev", 00:21:43.445 "uuid": "0f88b71f-399d-4bf6-a08b-abc8dfd92235", 00:21:43.445 "is_configured": true, 00:21:43.445 "data_offset": 2048, 00:21:43.445 "data_size": 63488 00:21:43.445 }, 00:21:43.445 { 00:21:43.445 "name": "BaseBdev2", 00:21:43.445 "uuid": "3d52c515-bac5-4b57-b22d-4fb4754b399f", 00:21:43.445 "is_configured": true, 00:21:43.445 "data_offset": 2048, 00:21:43.445 "data_size": 63488 00:21:43.445 }, 00:21:43.445 { 00:21:43.445 "name": "BaseBdev3", 00:21:43.445 "uuid": "1a372f20-1132-4d71-91f3-2611e13f28f7", 00:21:43.445 "is_configured": true, 00:21:43.445 "data_offset": 2048, 00:21:43.445 "data_size": 63488 00:21:43.445 } 00:21:43.445 ] 00:21:43.445 }' 00:21:43.445 20:20:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:43.445 20:20:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:44.011 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:21:44.011 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:21:44.011 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:21:44.011 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:21:44.011 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:21:44.011 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:21:44.011 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:21:44.011 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:44.011 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:44.011 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:21:44.011 [2024-10-01 20:20:39.122014] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:21:44.011 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:44.011 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:21:44.011 "name": "Existed_Raid", 00:21:44.011 "aliases": [ 00:21:44.011 "6d9ded77-d46e-4400-9cd9-609d47bdc4e6" 00:21:44.011 ], 00:21:44.011 "product_name": "Raid Volume", 00:21:44.011 "block_size": 512, 00:21:44.011 "num_blocks": 190464, 00:21:44.011 "uuid": "6d9ded77-d46e-4400-9cd9-609d47bdc4e6", 00:21:44.011 "assigned_rate_limits": { 00:21:44.011 "rw_ios_per_sec": 0, 00:21:44.011 "rw_mbytes_per_sec": 0, 00:21:44.011 "r_mbytes_per_sec": 0, 00:21:44.011 "w_mbytes_per_sec": 0 00:21:44.011 }, 00:21:44.011 "claimed": false, 00:21:44.011 "zoned": false, 00:21:44.011 "supported_io_types": { 00:21:44.011 "read": true, 00:21:44.011 "write": true, 00:21:44.011 "unmap": true, 00:21:44.011 "flush": true, 00:21:44.011 "reset": true, 00:21:44.011 "nvme_admin": false, 00:21:44.011 "nvme_io": false, 00:21:44.011 "nvme_io_md": false, 00:21:44.011 "write_zeroes": true, 00:21:44.011 "zcopy": false, 00:21:44.011 "get_zone_info": false, 00:21:44.011 "zone_management": false, 00:21:44.011 "zone_append": false, 00:21:44.011 "compare": false, 00:21:44.011 "compare_and_write": false, 00:21:44.011 "abort": false, 00:21:44.011 "seek_hole": false, 00:21:44.011 "seek_data": false, 00:21:44.011 "copy": false, 00:21:44.011 "nvme_iov_md": false 00:21:44.011 }, 00:21:44.011 "memory_domains": [ 00:21:44.011 { 00:21:44.011 "dma_device_id": "system", 00:21:44.011 "dma_device_type": 1 00:21:44.011 }, 00:21:44.011 { 00:21:44.011 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:44.011 "dma_device_type": 2 00:21:44.011 }, 00:21:44.011 { 00:21:44.011 "dma_device_id": "system", 00:21:44.011 "dma_device_type": 1 00:21:44.011 }, 00:21:44.012 { 00:21:44.012 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:44.012 "dma_device_type": 2 00:21:44.012 }, 00:21:44.012 { 00:21:44.012 "dma_device_id": "system", 00:21:44.012 "dma_device_type": 1 00:21:44.012 }, 00:21:44.012 { 00:21:44.012 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:44.012 "dma_device_type": 2 00:21:44.012 } 00:21:44.012 ], 00:21:44.012 "driver_specific": { 00:21:44.012 "raid": { 00:21:44.012 "uuid": "6d9ded77-d46e-4400-9cd9-609d47bdc4e6", 00:21:44.012 "strip_size_kb": 64, 00:21:44.012 "state": "online", 00:21:44.012 "raid_level": "raid0", 00:21:44.012 "superblock": true, 00:21:44.012 "num_base_bdevs": 3, 00:21:44.012 "num_base_bdevs_discovered": 3, 00:21:44.012 "num_base_bdevs_operational": 3, 00:21:44.012 "base_bdevs_list": [ 00:21:44.012 { 00:21:44.012 "name": "NewBaseBdev", 00:21:44.012 "uuid": "0f88b71f-399d-4bf6-a08b-abc8dfd92235", 00:21:44.012 "is_configured": true, 00:21:44.012 "data_offset": 2048, 00:21:44.012 "data_size": 63488 00:21:44.012 }, 00:21:44.012 { 00:21:44.012 "name": "BaseBdev2", 00:21:44.012 "uuid": "3d52c515-bac5-4b57-b22d-4fb4754b399f", 00:21:44.012 "is_configured": true, 00:21:44.012 "data_offset": 2048, 00:21:44.012 "data_size": 63488 00:21:44.012 }, 00:21:44.012 { 00:21:44.012 "name": "BaseBdev3", 00:21:44.012 "uuid": "1a372f20-1132-4d71-91f3-2611e13f28f7", 00:21:44.012 "is_configured": true, 00:21:44.012 "data_offset": 2048, 00:21:44.012 "data_size": 63488 00:21:44.012 } 00:21:44.012 ] 00:21:44.012 } 00:21:44.012 } 00:21:44.012 }' 00:21:44.012 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:21:44.012 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:21:44.012 BaseBdev2 00:21:44.012 BaseBdev3' 00:21:44.012 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:44.271 [2024-10-01 20:20:39.449658] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:21:44.271 [2024-10-01 20:20:39.449879] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:21:44.271 [2024-10-01 20:20:39.450025] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:21:44.271 [2024-10-01 20:20:39.450110] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:21:44.271 [2024-10-01 20:20:39.450132] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 64956 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 64956 ']' 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 64956 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 64956 00:21:44.271 killing process with pid 64956 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 64956' 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 64956 00:21:44.271 [2024-10-01 20:20:39.492539] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:21:44.271 20:20:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 64956 00:21:44.529 [2024-10-01 20:20:39.775782] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:21:46.433 20:20:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:21:46.433 00:21:46.433 real 0m13.052s 00:21:46.433 user 0m20.939s 00:21:46.433 sys 0m1.832s 00:21:46.433 ************************************ 00:21:46.433 END TEST raid_state_function_test_sb 00:21:46.433 ************************************ 00:21:46.433 20:20:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:21:46.433 20:20:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:21:46.433 20:20:41 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 3 00:21:46.433 20:20:41 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:21:46.433 20:20:41 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:21:46.433 20:20:41 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:21:46.433 ************************************ 00:21:46.433 START TEST raid_superblock_test 00:21:46.433 ************************************ 00:21:46.433 20:20:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid0 3 00:21:46.433 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:21:46.433 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:21:46.433 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:21:46.433 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:21:46.433 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=65604 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 65604 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 65604 ']' 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:46.434 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:21:46.434 20:20:41 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:46.692 [2024-10-01 20:20:41.785339] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:21:46.692 [2024-10-01 20:20:41.785816] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid65604 ] 00:21:46.951 [2024-10-01 20:20:41.961919] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:46.951 [2024-10-01 20:20:42.202897] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:21:47.210 [2024-10-01 20:20:42.405981] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:47.210 [2024-10-01 20:20:42.406042] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:47.779 malloc1 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:47.779 [2024-10-01 20:20:42.869827] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:21:47.779 [2024-10-01 20:20:42.870050] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:47.779 [2024-10-01 20:20:42.870134] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:21:47.779 [2024-10-01 20:20:42.870267] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:47.779 [2024-10-01 20:20:42.873350] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:47.779 [2024-10-01 20:20:42.873528] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:21:47.779 pt1 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:47.779 malloc2 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:47.779 [2024-10-01 20:20:42.922365] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:21:47.779 [2024-10-01 20:20:42.922443] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:47.779 [2024-10-01 20:20:42.922485] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:21:47.779 [2024-10-01 20:20:42.922503] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:47.779 [2024-10-01 20:20:42.925442] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:47.779 [2024-10-01 20:20:42.925491] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:21:47.779 pt2 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:47.779 malloc3 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:47.779 [2024-10-01 20:20:42.979064] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:21:47.779 [2024-10-01 20:20:42.979267] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:47.779 [2024-10-01 20:20:42.979326] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:21:47.779 [2024-10-01 20:20:42.979344] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:47.779 [2024-10-01 20:20:42.982330] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:47.779 [2024-10-01 20:20:42.982379] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:21:47.779 pt3 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:47.779 [2024-10-01 20:20:42.987307] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:21:47.779 [2024-10-01 20:20:42.990002] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:21:47.779 [2024-10-01 20:20:42.990105] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:21:47.779 [2024-10-01 20:20:42.990337] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:21:47.779 [2024-10-01 20:20:42.990362] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:21:47.779 [2024-10-01 20:20:42.990709] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:21:47.779 [2024-10-01 20:20:42.991129] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:21:47.779 [2024-10-01 20:20:42.991241] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:21:47.779 [2024-10-01 20:20:42.991694] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:47.779 20:20:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:47.779 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:48.039 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:48.039 "name": "raid_bdev1", 00:21:48.039 "uuid": "bebb83d3-95c5-4d29-a8b2-6e0c3d0362ba", 00:21:48.039 "strip_size_kb": 64, 00:21:48.039 "state": "online", 00:21:48.039 "raid_level": "raid0", 00:21:48.039 "superblock": true, 00:21:48.039 "num_base_bdevs": 3, 00:21:48.039 "num_base_bdevs_discovered": 3, 00:21:48.039 "num_base_bdevs_operational": 3, 00:21:48.039 "base_bdevs_list": [ 00:21:48.039 { 00:21:48.039 "name": "pt1", 00:21:48.039 "uuid": "00000000-0000-0000-0000-000000000001", 00:21:48.039 "is_configured": true, 00:21:48.039 "data_offset": 2048, 00:21:48.039 "data_size": 63488 00:21:48.039 }, 00:21:48.039 { 00:21:48.039 "name": "pt2", 00:21:48.039 "uuid": "00000000-0000-0000-0000-000000000002", 00:21:48.039 "is_configured": true, 00:21:48.039 "data_offset": 2048, 00:21:48.039 "data_size": 63488 00:21:48.039 }, 00:21:48.039 { 00:21:48.039 "name": "pt3", 00:21:48.039 "uuid": "00000000-0000-0000-0000-000000000003", 00:21:48.039 "is_configured": true, 00:21:48.039 "data_offset": 2048, 00:21:48.039 "data_size": 63488 00:21:48.039 } 00:21:48.039 ] 00:21:48.039 }' 00:21:48.039 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:48.040 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:48.299 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:21:48.299 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:21:48.299 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:21:48.299 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:21:48.299 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:21:48.299 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:21:48.299 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:21:48.300 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:48.300 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:48.300 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:21:48.300 [2024-10-01 20:20:43.516215] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:21:48.300 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:48.559 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:21:48.559 "name": "raid_bdev1", 00:21:48.559 "aliases": [ 00:21:48.559 "bebb83d3-95c5-4d29-a8b2-6e0c3d0362ba" 00:21:48.559 ], 00:21:48.559 "product_name": "Raid Volume", 00:21:48.559 "block_size": 512, 00:21:48.559 "num_blocks": 190464, 00:21:48.559 "uuid": "bebb83d3-95c5-4d29-a8b2-6e0c3d0362ba", 00:21:48.559 "assigned_rate_limits": { 00:21:48.559 "rw_ios_per_sec": 0, 00:21:48.559 "rw_mbytes_per_sec": 0, 00:21:48.559 "r_mbytes_per_sec": 0, 00:21:48.559 "w_mbytes_per_sec": 0 00:21:48.559 }, 00:21:48.559 "claimed": false, 00:21:48.559 "zoned": false, 00:21:48.559 "supported_io_types": { 00:21:48.559 "read": true, 00:21:48.559 "write": true, 00:21:48.559 "unmap": true, 00:21:48.559 "flush": true, 00:21:48.559 "reset": true, 00:21:48.559 "nvme_admin": false, 00:21:48.559 "nvme_io": false, 00:21:48.559 "nvme_io_md": false, 00:21:48.559 "write_zeroes": true, 00:21:48.559 "zcopy": false, 00:21:48.559 "get_zone_info": false, 00:21:48.559 "zone_management": false, 00:21:48.559 "zone_append": false, 00:21:48.559 "compare": false, 00:21:48.559 "compare_and_write": false, 00:21:48.559 "abort": false, 00:21:48.559 "seek_hole": false, 00:21:48.559 "seek_data": false, 00:21:48.559 "copy": false, 00:21:48.559 "nvme_iov_md": false 00:21:48.559 }, 00:21:48.559 "memory_domains": [ 00:21:48.559 { 00:21:48.559 "dma_device_id": "system", 00:21:48.559 "dma_device_type": 1 00:21:48.559 }, 00:21:48.559 { 00:21:48.559 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:48.559 "dma_device_type": 2 00:21:48.559 }, 00:21:48.559 { 00:21:48.559 "dma_device_id": "system", 00:21:48.559 "dma_device_type": 1 00:21:48.559 }, 00:21:48.559 { 00:21:48.559 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:48.559 "dma_device_type": 2 00:21:48.559 }, 00:21:48.559 { 00:21:48.559 "dma_device_id": "system", 00:21:48.559 "dma_device_type": 1 00:21:48.559 }, 00:21:48.559 { 00:21:48.559 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:48.559 "dma_device_type": 2 00:21:48.559 } 00:21:48.559 ], 00:21:48.559 "driver_specific": { 00:21:48.559 "raid": { 00:21:48.559 "uuid": "bebb83d3-95c5-4d29-a8b2-6e0c3d0362ba", 00:21:48.559 "strip_size_kb": 64, 00:21:48.559 "state": "online", 00:21:48.559 "raid_level": "raid0", 00:21:48.559 "superblock": true, 00:21:48.559 "num_base_bdevs": 3, 00:21:48.559 "num_base_bdevs_discovered": 3, 00:21:48.559 "num_base_bdevs_operational": 3, 00:21:48.559 "base_bdevs_list": [ 00:21:48.559 { 00:21:48.559 "name": "pt1", 00:21:48.559 "uuid": "00000000-0000-0000-0000-000000000001", 00:21:48.559 "is_configured": true, 00:21:48.559 "data_offset": 2048, 00:21:48.559 "data_size": 63488 00:21:48.559 }, 00:21:48.559 { 00:21:48.559 "name": "pt2", 00:21:48.559 "uuid": "00000000-0000-0000-0000-000000000002", 00:21:48.559 "is_configured": true, 00:21:48.559 "data_offset": 2048, 00:21:48.559 "data_size": 63488 00:21:48.559 }, 00:21:48.559 { 00:21:48.559 "name": "pt3", 00:21:48.559 "uuid": "00000000-0000-0000-0000-000000000003", 00:21:48.559 "is_configured": true, 00:21:48.559 "data_offset": 2048, 00:21:48.559 "data_size": 63488 00:21:48.559 } 00:21:48.559 ] 00:21:48.559 } 00:21:48.559 } 00:21:48.559 }' 00:21:48.559 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:21:48.559 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:21:48.559 pt2 00:21:48.559 pt3' 00:21:48.559 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:48.559 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:21:48.559 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:48.559 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:21:48.559 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:48.559 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:48.559 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:48.559 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:48.559 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:48.559 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:48.560 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:48.560 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:21:48.560 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:48.560 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:48.560 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:48.560 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:48.560 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:48.560 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:48.560 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:48.560 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:21:48.560 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:48.560 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:48.560 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:48.560 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:48.819 [2024-10-01 20:20:43.860248] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=bebb83d3-95c5-4d29-a8b2-6e0c3d0362ba 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z bebb83d3-95c5-4d29-a8b2-6e0c3d0362ba ']' 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:48.819 [2024-10-01 20:20:43.907888] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:21:48.819 [2024-10-01 20:20:43.908081] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:21:48.819 [2024-10-01 20:20:43.908298] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:21:48.819 [2024-10-01 20:20:43.908491] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:21:48.819 [2024-10-01 20:20:43.908645] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:48.819 20:20:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:21:48.819 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:48.819 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:21:48.819 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:21:48.819 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:21:48.819 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:21:48.819 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:21:48.819 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:21:48.819 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:21:48.820 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:21:48.820 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:21:48.820 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:48.820 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:48.820 [2024-10-01 20:20:44.056007] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:21:48.820 [2024-10-01 20:20:44.058857] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:21:48.820 [2024-10-01 20:20:44.059062] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:21:48.820 [2024-10-01 20:20:44.059168] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:21:48.820 [2024-10-01 20:20:44.059244] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:21:48.820 [2024-10-01 20:20:44.059280] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:21:48.820 [2024-10-01 20:20:44.059309] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:21:48.820 [2024-10-01 20:20:44.059323] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:21:48.820 request: 00:21:48.820 { 00:21:48.820 "name": "raid_bdev1", 00:21:48.820 "raid_level": "raid0", 00:21:48.820 "base_bdevs": [ 00:21:48.820 "malloc1", 00:21:48.820 "malloc2", 00:21:48.820 "malloc3" 00:21:48.820 ], 00:21:48.820 "strip_size_kb": 64, 00:21:48.820 "superblock": false, 00:21:48.820 "method": "bdev_raid_create", 00:21:48.820 "req_id": 1 00:21:48.820 } 00:21:48.820 Got JSON-RPC error response 00:21:48.820 response: 00:21:48.820 { 00:21:48.820 "code": -17, 00:21:48.820 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:21:48.820 } 00:21:48.820 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:21:48.820 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:21:48.820 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:21:48.820 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:21:48.820 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:21:48.820 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:48.820 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:48.820 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:48.820 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:49.079 [2024-10-01 20:20:44.140159] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:21:49.079 [2024-10-01 20:20:44.140412] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:49.079 [2024-10-01 20:20:44.140493] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:21:49.079 [2024-10-01 20:20:44.140660] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:49.079 [2024-10-01 20:20:44.143811] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:49.079 [2024-10-01 20:20:44.143878] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:21:49.079 [2024-10-01 20:20:44.144011] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:21:49.079 [2024-10-01 20:20:44.144079] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:21:49.079 pt1 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:49.079 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:49.079 "name": "raid_bdev1", 00:21:49.079 "uuid": "bebb83d3-95c5-4d29-a8b2-6e0c3d0362ba", 00:21:49.079 "strip_size_kb": 64, 00:21:49.079 "state": "configuring", 00:21:49.079 "raid_level": "raid0", 00:21:49.079 "superblock": true, 00:21:49.080 "num_base_bdevs": 3, 00:21:49.080 "num_base_bdevs_discovered": 1, 00:21:49.080 "num_base_bdevs_operational": 3, 00:21:49.080 "base_bdevs_list": [ 00:21:49.080 { 00:21:49.080 "name": "pt1", 00:21:49.080 "uuid": "00000000-0000-0000-0000-000000000001", 00:21:49.080 "is_configured": true, 00:21:49.080 "data_offset": 2048, 00:21:49.080 "data_size": 63488 00:21:49.080 }, 00:21:49.080 { 00:21:49.080 "name": null, 00:21:49.080 "uuid": "00000000-0000-0000-0000-000000000002", 00:21:49.080 "is_configured": false, 00:21:49.080 "data_offset": 2048, 00:21:49.080 "data_size": 63488 00:21:49.080 }, 00:21:49.080 { 00:21:49.080 "name": null, 00:21:49.080 "uuid": "00000000-0000-0000-0000-000000000003", 00:21:49.080 "is_configured": false, 00:21:49.080 "data_offset": 2048, 00:21:49.080 "data_size": 63488 00:21:49.080 } 00:21:49.080 ] 00:21:49.080 }' 00:21:49.080 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:49.080 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:49.647 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:49.648 [2024-10-01 20:20:44.672568] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:21:49.648 [2024-10-01 20:20:44.672901] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:49.648 [2024-10-01 20:20:44.672961] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:21:49.648 [2024-10-01 20:20:44.672982] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:49.648 [2024-10-01 20:20:44.673613] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:49.648 [2024-10-01 20:20:44.673652] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:21:49.648 [2024-10-01 20:20:44.673805] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:21:49.648 [2024-10-01 20:20:44.673841] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:21:49.648 pt2 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:49.648 [2024-10-01 20:20:44.684589] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:49.648 "name": "raid_bdev1", 00:21:49.648 "uuid": "bebb83d3-95c5-4d29-a8b2-6e0c3d0362ba", 00:21:49.648 "strip_size_kb": 64, 00:21:49.648 "state": "configuring", 00:21:49.648 "raid_level": "raid0", 00:21:49.648 "superblock": true, 00:21:49.648 "num_base_bdevs": 3, 00:21:49.648 "num_base_bdevs_discovered": 1, 00:21:49.648 "num_base_bdevs_operational": 3, 00:21:49.648 "base_bdevs_list": [ 00:21:49.648 { 00:21:49.648 "name": "pt1", 00:21:49.648 "uuid": "00000000-0000-0000-0000-000000000001", 00:21:49.648 "is_configured": true, 00:21:49.648 "data_offset": 2048, 00:21:49.648 "data_size": 63488 00:21:49.648 }, 00:21:49.648 { 00:21:49.648 "name": null, 00:21:49.648 "uuid": "00000000-0000-0000-0000-000000000002", 00:21:49.648 "is_configured": false, 00:21:49.648 "data_offset": 0, 00:21:49.648 "data_size": 63488 00:21:49.648 }, 00:21:49.648 { 00:21:49.648 "name": null, 00:21:49.648 "uuid": "00000000-0000-0000-0000-000000000003", 00:21:49.648 "is_configured": false, 00:21:49.648 "data_offset": 2048, 00:21:49.648 "data_size": 63488 00:21:49.648 } 00:21:49.648 ] 00:21:49.648 }' 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:49.648 20:20:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:50.214 [2024-10-01 20:20:45.208644] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:21:50.214 [2024-10-01 20:20:45.208919] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:50.214 [2024-10-01 20:20:45.209076] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:21:50.214 [2024-10-01 20:20:45.209205] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:50.214 [2024-10-01 20:20:45.210015] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:50.214 [2024-10-01 20:20:45.210056] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:21:50.214 [2024-10-01 20:20:45.210169] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:21:50.214 [2024-10-01 20:20:45.210223] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:21:50.214 pt2 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:50.214 [2024-10-01 20:20:45.216614] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:21:50.214 [2024-10-01 20:20:45.216822] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:50.214 [2024-10-01 20:20:45.216911] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:21:50.214 [2024-10-01 20:20:45.217157] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:50.214 [2024-10-01 20:20:45.217753] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:50.214 [2024-10-01 20:20:45.217921] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:21:50.214 [2024-10-01 20:20:45.218129] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:21:50.214 [2024-10-01 20:20:45.218301] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:21:50.214 [2024-10-01 20:20:45.218573] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:21:50.214 [2024-10-01 20:20:45.218708] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:21:50.214 [2024-10-01 20:20:45.219165] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:21:50.214 [2024-10-01 20:20:45.219516] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:21:50.214 [2024-10-01 20:20:45.219666] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:21:50.214 [2024-10-01 20:20:45.220074] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:50.214 pt3 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:50.214 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:50.214 "name": "raid_bdev1", 00:21:50.214 "uuid": "bebb83d3-95c5-4d29-a8b2-6e0c3d0362ba", 00:21:50.214 "strip_size_kb": 64, 00:21:50.214 "state": "online", 00:21:50.214 "raid_level": "raid0", 00:21:50.214 "superblock": true, 00:21:50.214 "num_base_bdevs": 3, 00:21:50.214 "num_base_bdevs_discovered": 3, 00:21:50.214 "num_base_bdevs_operational": 3, 00:21:50.214 "base_bdevs_list": [ 00:21:50.214 { 00:21:50.215 "name": "pt1", 00:21:50.215 "uuid": "00000000-0000-0000-0000-000000000001", 00:21:50.215 "is_configured": true, 00:21:50.215 "data_offset": 2048, 00:21:50.215 "data_size": 63488 00:21:50.215 }, 00:21:50.215 { 00:21:50.215 "name": "pt2", 00:21:50.215 "uuid": "00000000-0000-0000-0000-000000000002", 00:21:50.215 "is_configured": true, 00:21:50.215 "data_offset": 2048, 00:21:50.215 "data_size": 63488 00:21:50.215 }, 00:21:50.215 { 00:21:50.215 "name": "pt3", 00:21:50.215 "uuid": "00000000-0000-0000-0000-000000000003", 00:21:50.215 "is_configured": true, 00:21:50.215 "data_offset": 2048, 00:21:50.215 "data_size": 63488 00:21:50.215 } 00:21:50.215 ] 00:21:50.215 }' 00:21:50.215 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:50.215 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:50.781 [2024-10-01 20:20:45.737189] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:21:50.781 "name": "raid_bdev1", 00:21:50.781 "aliases": [ 00:21:50.781 "bebb83d3-95c5-4d29-a8b2-6e0c3d0362ba" 00:21:50.781 ], 00:21:50.781 "product_name": "Raid Volume", 00:21:50.781 "block_size": 512, 00:21:50.781 "num_blocks": 190464, 00:21:50.781 "uuid": "bebb83d3-95c5-4d29-a8b2-6e0c3d0362ba", 00:21:50.781 "assigned_rate_limits": { 00:21:50.781 "rw_ios_per_sec": 0, 00:21:50.781 "rw_mbytes_per_sec": 0, 00:21:50.781 "r_mbytes_per_sec": 0, 00:21:50.781 "w_mbytes_per_sec": 0 00:21:50.781 }, 00:21:50.781 "claimed": false, 00:21:50.781 "zoned": false, 00:21:50.781 "supported_io_types": { 00:21:50.781 "read": true, 00:21:50.781 "write": true, 00:21:50.781 "unmap": true, 00:21:50.781 "flush": true, 00:21:50.781 "reset": true, 00:21:50.781 "nvme_admin": false, 00:21:50.781 "nvme_io": false, 00:21:50.781 "nvme_io_md": false, 00:21:50.781 "write_zeroes": true, 00:21:50.781 "zcopy": false, 00:21:50.781 "get_zone_info": false, 00:21:50.781 "zone_management": false, 00:21:50.781 "zone_append": false, 00:21:50.781 "compare": false, 00:21:50.781 "compare_and_write": false, 00:21:50.781 "abort": false, 00:21:50.781 "seek_hole": false, 00:21:50.781 "seek_data": false, 00:21:50.781 "copy": false, 00:21:50.781 "nvme_iov_md": false 00:21:50.781 }, 00:21:50.781 "memory_domains": [ 00:21:50.781 { 00:21:50.781 "dma_device_id": "system", 00:21:50.781 "dma_device_type": 1 00:21:50.781 }, 00:21:50.781 { 00:21:50.781 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:50.781 "dma_device_type": 2 00:21:50.781 }, 00:21:50.781 { 00:21:50.781 "dma_device_id": "system", 00:21:50.781 "dma_device_type": 1 00:21:50.781 }, 00:21:50.781 { 00:21:50.781 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:50.781 "dma_device_type": 2 00:21:50.781 }, 00:21:50.781 { 00:21:50.781 "dma_device_id": "system", 00:21:50.781 "dma_device_type": 1 00:21:50.781 }, 00:21:50.781 { 00:21:50.781 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:21:50.781 "dma_device_type": 2 00:21:50.781 } 00:21:50.781 ], 00:21:50.781 "driver_specific": { 00:21:50.781 "raid": { 00:21:50.781 "uuid": "bebb83d3-95c5-4d29-a8b2-6e0c3d0362ba", 00:21:50.781 "strip_size_kb": 64, 00:21:50.781 "state": "online", 00:21:50.781 "raid_level": "raid0", 00:21:50.781 "superblock": true, 00:21:50.781 "num_base_bdevs": 3, 00:21:50.781 "num_base_bdevs_discovered": 3, 00:21:50.781 "num_base_bdevs_operational": 3, 00:21:50.781 "base_bdevs_list": [ 00:21:50.781 { 00:21:50.781 "name": "pt1", 00:21:50.781 "uuid": "00000000-0000-0000-0000-000000000001", 00:21:50.781 "is_configured": true, 00:21:50.781 "data_offset": 2048, 00:21:50.781 "data_size": 63488 00:21:50.781 }, 00:21:50.781 { 00:21:50.781 "name": "pt2", 00:21:50.781 "uuid": "00000000-0000-0000-0000-000000000002", 00:21:50.781 "is_configured": true, 00:21:50.781 "data_offset": 2048, 00:21:50.781 "data_size": 63488 00:21:50.781 }, 00:21:50.781 { 00:21:50.781 "name": "pt3", 00:21:50.781 "uuid": "00000000-0000-0000-0000-000000000003", 00:21:50.781 "is_configured": true, 00:21:50.781 "data_offset": 2048, 00:21:50.781 "data_size": 63488 00:21:50.781 } 00:21:50.781 ] 00:21:50.781 } 00:21:50.781 } 00:21:50.781 }' 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:21:50.781 pt2 00:21:50.781 pt3' 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:50.781 20:20:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:50.781 20:20:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:50.781 20:20:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:50.781 20:20:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:21:50.781 20:20:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:21:50.781 20:20:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:21:50.781 20:20:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:50.781 20:20:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:50.781 20:20:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:21:51.039 [2024-10-01 20:20:46.065227] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' bebb83d3-95c5-4d29-a8b2-6e0c3d0362ba '!=' bebb83d3-95c5-4d29-a8b2-6e0c3d0362ba ']' 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 65604 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 65604 ']' 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 65604 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 65604 00:21:51.039 killing process with pid 65604 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 65604' 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 65604 00:21:51.039 [2024-10-01 20:20:46.146972] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:21:51.039 20:20:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 65604 00:21:51.039 [2024-10-01 20:20:46.147123] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:21:51.039 [2024-10-01 20:20:46.147206] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:21:51.039 [2024-10-01 20:20:46.147228] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:21:51.297 [2024-10-01 20:20:46.418840] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:21:53.198 20:20:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:21:53.198 00:21:53.198 real 0m6.547s 00:21:53.198 user 0m9.355s 00:21:53.198 sys 0m0.976s 00:21:53.198 20:20:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:21:53.198 ************************************ 00:21:53.198 END TEST raid_superblock_test 00:21:53.198 ************************************ 00:21:53.198 20:20:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:21:53.198 20:20:48 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 3 read 00:21:53.198 20:20:48 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:21:53.198 20:20:48 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:21:53.198 20:20:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:21:53.198 ************************************ 00:21:53.198 START TEST raid_read_error_test 00:21:53.198 ************************************ 00:21:53.198 20:20:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 3 read 00:21:53.198 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:21:53.198 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:21:53.198 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:21:53.198 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:21:53.198 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:21:53.198 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:21:53.198 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:21:53.198 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:21:53.198 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:21:53.198 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:21:53.198 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:21:53.199 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.v9oJAfJtpj 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=65868 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 65868 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 65868 ']' 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:21:53.199 20:20:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:53.199 [2024-10-01 20:20:48.420357] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:21:53.199 [2024-10-01 20:20:48.420546] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid65868 ] 00:21:53.457 [2024-10-01 20:20:48.597614] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:53.715 [2024-10-01 20:20:48.841472] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:21:54.059 [2024-10-01 20:20:49.045433] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:54.059 [2024-10-01 20:20:49.045535] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:54.319 BaseBdev1_malloc 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:54.319 true 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:54.319 [2024-10-01 20:20:49.508532] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:21:54.319 [2024-10-01 20:20:49.508608] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:54.319 [2024-10-01 20:20:49.508638] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:21:54.319 [2024-10-01 20:20:49.508658] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:54.319 [2024-10-01 20:20:49.511741] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:54.319 [2024-10-01 20:20:49.511908] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:21:54.319 BaseBdev1 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:54.319 BaseBdev2_malloc 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:54.319 true 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:54.319 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:54.319 [2024-10-01 20:20:49.569593] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:21:54.319 [2024-10-01 20:20:49.569659] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:54.319 [2024-10-01 20:20:49.569764] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:21:54.319 [2024-10-01 20:20:49.569897] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:54.579 [2024-10-01 20:20:49.572872] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:54.579 [2024-10-01 20:20:49.573045] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:21:54.579 BaseBdev2 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:54.579 BaseBdev3_malloc 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:54.579 true 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:54.579 [2024-10-01 20:20:49.631328] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:21:54.579 [2024-10-01 20:20:49.631558] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:21:54.579 [2024-10-01 20:20:49.631608] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:21:54.579 [2024-10-01 20:20:49.631628] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:21:54.579 [2024-10-01 20:20:49.634566] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:21:54.579 [2024-10-01 20:20:49.634619] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:21:54.579 BaseBdev3 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:54.579 [2024-10-01 20:20:49.639570] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:21:54.579 [2024-10-01 20:20:49.642212] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:21:54.579 [2024-10-01 20:20:49.642338] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:21:54.579 [2024-10-01 20:20:49.642645] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:21:54.579 [2024-10-01 20:20:49.642667] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:21:54.579 [2024-10-01 20:20:49.643044] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:21:54.579 [2024-10-01 20:20:49.643252] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:21:54.579 [2024-10-01 20:20:49.643274] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:21:54.579 [2024-10-01 20:20:49.643520] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:54.579 "name": "raid_bdev1", 00:21:54.579 "uuid": "afeeef2a-bd61-4d4f-9bb7-7208b31f2bce", 00:21:54.579 "strip_size_kb": 64, 00:21:54.579 "state": "online", 00:21:54.579 "raid_level": "raid0", 00:21:54.579 "superblock": true, 00:21:54.579 "num_base_bdevs": 3, 00:21:54.579 "num_base_bdevs_discovered": 3, 00:21:54.579 "num_base_bdevs_operational": 3, 00:21:54.579 "base_bdevs_list": [ 00:21:54.579 { 00:21:54.579 "name": "BaseBdev1", 00:21:54.579 "uuid": "22ff0ce0-ec65-5735-9862-0b8ceb35f4f6", 00:21:54.579 "is_configured": true, 00:21:54.579 "data_offset": 2048, 00:21:54.579 "data_size": 63488 00:21:54.579 }, 00:21:54.579 { 00:21:54.579 "name": "BaseBdev2", 00:21:54.579 "uuid": "53adfdf4-2d69-5dd5-b39a-5a3ed65cb8b5", 00:21:54.579 "is_configured": true, 00:21:54.579 "data_offset": 2048, 00:21:54.579 "data_size": 63488 00:21:54.579 }, 00:21:54.579 { 00:21:54.579 "name": "BaseBdev3", 00:21:54.579 "uuid": "60ea8887-55d7-55f6-8aba-236e9a2570dd", 00:21:54.579 "is_configured": true, 00:21:54.579 "data_offset": 2048, 00:21:54.579 "data_size": 63488 00:21:54.579 } 00:21:54.579 ] 00:21:54.579 }' 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:54.579 20:20:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:55.145 20:20:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:21:55.145 20:20:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:21:55.145 [2024-10-01 20:20:50.265212] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:21:56.080 "name": "raid_bdev1", 00:21:56.080 "uuid": "afeeef2a-bd61-4d4f-9bb7-7208b31f2bce", 00:21:56.080 "strip_size_kb": 64, 00:21:56.080 "state": "online", 00:21:56.080 "raid_level": "raid0", 00:21:56.080 "superblock": true, 00:21:56.080 "num_base_bdevs": 3, 00:21:56.080 "num_base_bdevs_discovered": 3, 00:21:56.080 "num_base_bdevs_operational": 3, 00:21:56.080 "base_bdevs_list": [ 00:21:56.080 { 00:21:56.080 "name": "BaseBdev1", 00:21:56.080 "uuid": "22ff0ce0-ec65-5735-9862-0b8ceb35f4f6", 00:21:56.080 "is_configured": true, 00:21:56.080 "data_offset": 2048, 00:21:56.080 "data_size": 63488 00:21:56.080 }, 00:21:56.080 { 00:21:56.080 "name": "BaseBdev2", 00:21:56.080 "uuid": "53adfdf4-2d69-5dd5-b39a-5a3ed65cb8b5", 00:21:56.080 "is_configured": true, 00:21:56.080 "data_offset": 2048, 00:21:56.080 "data_size": 63488 00:21:56.080 }, 00:21:56.080 { 00:21:56.080 "name": "BaseBdev3", 00:21:56.080 "uuid": "60ea8887-55d7-55f6-8aba-236e9a2570dd", 00:21:56.080 "is_configured": true, 00:21:56.080 "data_offset": 2048, 00:21:56.080 "data_size": 63488 00:21:56.080 } 00:21:56.080 ] 00:21:56.080 }' 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:21:56.080 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:56.646 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:21:56.646 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:56.646 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:56.646 [2024-10-01 20:20:51.737710] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:21:56.646 [2024-10-01 20:20:51.737762] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:21:56.646 [2024-10-01 20:20:51.741086] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:21:56.646 [2024-10-01 20:20:51.741155] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:21:56.646 [2024-10-01 20:20:51.741211] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:21:56.646 [2024-10-01 20:20:51.741227] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:21:56.646 { 00:21:56.646 "results": [ 00:21:56.646 { 00:21:56.646 "job": "raid_bdev1", 00:21:56.646 "core_mask": "0x1", 00:21:56.646 "workload": "randrw", 00:21:56.646 "percentage": 50, 00:21:56.646 "status": "finished", 00:21:56.646 "queue_depth": 1, 00:21:56.646 "io_size": 131072, 00:21:56.646 "runtime": 1.469975, 00:21:56.646 "iops": 10251.194748210004, 00:21:56.646 "mibps": 1281.3993435262505, 00:21:56.646 "io_failed": 1, 00:21:56.646 "io_timeout": 0, 00:21:56.646 "avg_latency_us": 136.03195560113411, 00:21:56.646 "min_latency_us": 42.82181818181818, 00:21:56.646 "max_latency_us": 1861.8181818181818 00:21:56.646 } 00:21:56.646 ], 00:21:56.646 "core_count": 1 00:21:56.646 } 00:21:56.646 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:56.646 20:20:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 65868 00:21:56.646 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 65868 ']' 00:21:56.646 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 65868 00:21:56.646 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:21:56.646 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:21:56.646 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 65868 00:21:56.646 killing process with pid 65868 00:21:56.646 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:21:56.646 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:21:56.646 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 65868' 00:21:56.646 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 65868 00:21:56.646 [2024-10-01 20:20:51.776701] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:21:56.646 20:20:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 65868 00:21:56.904 [2024-10-01 20:20:51.988987] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:21:58.806 20:20:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.v9oJAfJtpj 00:21:58.806 20:20:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:21:58.806 20:20:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:21:58.806 20:20:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.68 00:21:58.806 20:20:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:21:58.806 20:20:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:21:58.806 20:20:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:21:58.806 20:20:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.68 != \0\.\0\0 ]] 00:21:58.806 00:21:58.806 real 0m5.596s 00:21:58.806 user 0m6.722s 00:21:58.806 sys 0m0.692s 00:21:58.806 20:20:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:21:58.806 ************************************ 00:21:58.806 END TEST raid_read_error_test 00:21:58.806 ************************************ 00:21:58.806 20:20:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:58.806 20:20:53 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 3 write 00:21:58.806 20:20:53 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:21:58.806 20:20:53 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:21:58.806 20:20:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:21:58.806 ************************************ 00:21:58.806 START TEST raid_write_error_test 00:21:58.806 ************************************ 00:21:58.806 20:20:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 3 write 00:21:58.806 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.3g5EncgBR2 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=66025 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 66025 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 66025 ']' 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:21:58.807 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:21:58.807 20:20:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:21:58.807 [2024-10-01 20:20:54.042858] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:21:58.807 [2024-10-01 20:20:54.043052] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid66025 ] 00:21:59.065 [2024-10-01 20:20:54.217612] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:59.323 [2024-10-01 20:20:54.457568] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:21:59.582 [2024-10-01 20:20:54.660596] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:59.582 [2024-10-01 20:20:54.660651] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:21:59.840 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:21:59.840 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:21:59.840 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:21:59.840 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:21:59.840 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:59.840 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:00.099 BaseBdev1_malloc 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:00.099 true 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:00.099 [2024-10-01 20:20:55.129455] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:22:00.099 [2024-10-01 20:20:55.129538] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:00.099 [2024-10-01 20:20:55.129566] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:22:00.099 [2024-10-01 20:20:55.129584] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:00.099 [2024-10-01 20:20:55.132545] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:00.099 [2024-10-01 20:20:55.132607] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:22:00.099 BaseBdev1 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:00.099 BaseBdev2_malloc 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:00.099 true 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:00.099 [2024-10-01 20:20:55.190424] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:22:00.099 [2024-10-01 20:20:55.190511] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:00.099 [2024-10-01 20:20:55.190539] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:22:00.099 [2024-10-01 20:20:55.190558] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:00.099 [2024-10-01 20:20:55.193451] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:00.099 [2024-10-01 20:20:55.193515] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:22:00.099 BaseBdev2 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:00.099 BaseBdev3_malloc 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:00.099 true 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:00.099 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:00.099 [2024-10-01 20:20:55.247041] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:22:00.099 [2024-10-01 20:20:55.247110] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:00.099 [2024-10-01 20:20:55.247138] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:22:00.099 [2024-10-01 20:20:55.247157] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:00.099 [2024-10-01 20:20:55.250101] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:00.099 [2024-10-01 20:20:55.250150] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:22:00.099 BaseBdev3 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:00.100 [2024-10-01 20:20:55.255136] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:22:00.100 [2024-10-01 20:20:55.257644] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:22:00.100 [2024-10-01 20:20:55.257786] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:22:00.100 [2024-10-01 20:20:55.258064] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:22:00.100 [2024-10-01 20:20:55.258083] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:22:00.100 [2024-10-01 20:20:55.258411] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:22:00.100 [2024-10-01 20:20:55.258619] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:22:00.100 [2024-10-01 20:20:55.258640] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:22:00.100 [2024-10-01 20:20:55.258847] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:00.100 "name": "raid_bdev1", 00:22:00.100 "uuid": "56afb091-d371-4245-8d9f-00bc80ee07c8", 00:22:00.100 "strip_size_kb": 64, 00:22:00.100 "state": "online", 00:22:00.100 "raid_level": "raid0", 00:22:00.100 "superblock": true, 00:22:00.100 "num_base_bdevs": 3, 00:22:00.100 "num_base_bdevs_discovered": 3, 00:22:00.100 "num_base_bdevs_operational": 3, 00:22:00.100 "base_bdevs_list": [ 00:22:00.100 { 00:22:00.100 "name": "BaseBdev1", 00:22:00.100 "uuid": "629d9188-4c64-5904-8507-6ea03c6b6a67", 00:22:00.100 "is_configured": true, 00:22:00.100 "data_offset": 2048, 00:22:00.100 "data_size": 63488 00:22:00.100 }, 00:22:00.100 { 00:22:00.100 "name": "BaseBdev2", 00:22:00.100 "uuid": "49223c09-6668-503a-94c4-147d051bbb90", 00:22:00.100 "is_configured": true, 00:22:00.100 "data_offset": 2048, 00:22:00.100 "data_size": 63488 00:22:00.100 }, 00:22:00.100 { 00:22:00.100 "name": "BaseBdev3", 00:22:00.100 "uuid": "96238077-2a13-5043-846e-d97430eba64d", 00:22:00.100 "is_configured": true, 00:22:00.100 "data_offset": 2048, 00:22:00.100 "data_size": 63488 00:22:00.100 } 00:22:00.100 ] 00:22:00.100 }' 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:00.100 20:20:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:00.666 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:22:00.666 20:20:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:22:00.666 [2024-10-01 20:20:55.904700] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:01.601 20:20:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:01.859 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:01.859 "name": "raid_bdev1", 00:22:01.859 "uuid": "56afb091-d371-4245-8d9f-00bc80ee07c8", 00:22:01.859 "strip_size_kb": 64, 00:22:01.859 "state": "online", 00:22:01.859 "raid_level": "raid0", 00:22:01.859 "superblock": true, 00:22:01.859 "num_base_bdevs": 3, 00:22:01.859 "num_base_bdevs_discovered": 3, 00:22:01.859 "num_base_bdevs_operational": 3, 00:22:01.859 "base_bdevs_list": [ 00:22:01.859 { 00:22:01.859 "name": "BaseBdev1", 00:22:01.859 "uuid": "629d9188-4c64-5904-8507-6ea03c6b6a67", 00:22:01.859 "is_configured": true, 00:22:01.859 "data_offset": 2048, 00:22:01.859 "data_size": 63488 00:22:01.859 }, 00:22:01.859 { 00:22:01.859 "name": "BaseBdev2", 00:22:01.859 "uuid": "49223c09-6668-503a-94c4-147d051bbb90", 00:22:01.859 "is_configured": true, 00:22:01.859 "data_offset": 2048, 00:22:01.859 "data_size": 63488 00:22:01.859 }, 00:22:01.859 { 00:22:01.859 "name": "BaseBdev3", 00:22:01.859 "uuid": "96238077-2a13-5043-846e-d97430eba64d", 00:22:01.859 "is_configured": true, 00:22:01.859 "data_offset": 2048, 00:22:01.859 "data_size": 63488 00:22:01.859 } 00:22:01.859 ] 00:22:01.859 }' 00:22:01.859 20:20:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:01.859 20:20:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:02.118 20:20:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:22:02.118 20:20:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:02.118 20:20:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:02.118 [2024-10-01 20:20:57.328392] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:22:02.118 [2024-10-01 20:20:57.328433] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:22:02.118 [2024-10-01 20:20:57.331805] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:02.118 [2024-10-01 20:20:57.331867] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:02.118 [2024-10-01 20:20:57.331921] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:22:02.118 [2024-10-01 20:20:57.331936] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:22:02.118 { 00:22:02.118 "results": [ 00:22:02.118 { 00:22:02.118 "job": "raid_bdev1", 00:22:02.118 "core_mask": "0x1", 00:22:02.118 "workload": "randrw", 00:22:02.118 "percentage": 50, 00:22:02.118 "status": "finished", 00:22:02.118 "queue_depth": 1, 00:22:02.118 "io_size": 131072, 00:22:02.118 "runtime": 1.421035, 00:22:02.118 "iops": 10414.24032483366, 00:22:02.118 "mibps": 1301.7800406042074, 00:22:02.118 "io_failed": 1, 00:22:02.118 "io_timeout": 0, 00:22:02.118 "avg_latency_us": 133.90102407862406, 00:22:02.118 "min_latency_us": 28.85818181818182, 00:22:02.118 "max_latency_us": 1854.370909090909 00:22:02.118 } 00:22:02.118 ], 00:22:02.118 "core_count": 1 00:22:02.118 } 00:22:02.118 20:20:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:02.118 20:20:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 66025 00:22:02.118 20:20:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 66025 ']' 00:22:02.118 20:20:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 66025 00:22:02.118 20:20:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:22:02.118 20:20:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:02.118 20:20:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 66025 00:22:02.118 20:20:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:22:02.118 killing process with pid 66025 00:22:02.118 20:20:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:22:02.118 20:20:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 66025' 00:22:02.118 20:20:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 66025 00:22:02.118 [2024-10-01 20:20:57.365891] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:22:02.118 20:20:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 66025 00:22:02.377 [2024-10-01 20:20:57.572421] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:22:04.313 20:20:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.3g5EncgBR2 00:22:04.313 20:20:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:22:04.313 20:20:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:22:04.314 20:20:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.70 00:22:04.314 20:20:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:22:04.314 20:20:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:22:04.314 20:20:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:22:04.314 20:20:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.70 != \0\.\0\0 ]] 00:22:04.314 00:22:04.314 real 0m5.501s 00:22:04.314 user 0m6.570s 00:22:04.314 sys 0m0.680s 00:22:04.314 20:20:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:22:04.314 ************************************ 00:22:04.314 END TEST raid_write_error_test 00:22:04.314 ************************************ 00:22:04.314 20:20:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:04.314 20:20:59 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:22:04.314 20:20:59 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 3 false 00:22:04.314 20:20:59 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:22:04.314 20:20:59 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:22:04.314 20:20:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:22:04.314 ************************************ 00:22:04.314 START TEST raid_state_function_test 00:22:04.314 ************************************ 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 3 false 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=66174 00:22:04.314 Process raid pid: 66174 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 66174' 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 66174 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 66174 ']' 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:04.314 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:04.314 20:20:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:04.572 [2024-10-01 20:20:59.582250] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:22:04.572 [2024-10-01 20:20:59.582428] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:22:04.572 [2024-10-01 20:20:59.749043] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:04.830 [2024-10-01 20:20:59.999119] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:22:05.089 [2024-10-01 20:21:00.211973] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:05.089 [2024-10-01 20:21:00.212059] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:05.656 [2024-10-01 20:21:00.628050] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:22:05.656 [2024-10-01 20:21:00.628115] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:22:05.656 [2024-10-01 20:21:00.628132] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:22:05.656 [2024-10-01 20:21:00.628152] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:22:05.656 [2024-10-01 20:21:00.628162] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:22:05.656 [2024-10-01 20:21:00.628179] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:05.656 "name": "Existed_Raid", 00:22:05.656 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:05.656 "strip_size_kb": 64, 00:22:05.656 "state": "configuring", 00:22:05.656 "raid_level": "concat", 00:22:05.656 "superblock": false, 00:22:05.656 "num_base_bdevs": 3, 00:22:05.656 "num_base_bdevs_discovered": 0, 00:22:05.656 "num_base_bdevs_operational": 3, 00:22:05.656 "base_bdevs_list": [ 00:22:05.656 { 00:22:05.656 "name": "BaseBdev1", 00:22:05.656 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:05.656 "is_configured": false, 00:22:05.656 "data_offset": 0, 00:22:05.656 "data_size": 0 00:22:05.656 }, 00:22:05.656 { 00:22:05.656 "name": "BaseBdev2", 00:22:05.656 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:05.656 "is_configured": false, 00:22:05.656 "data_offset": 0, 00:22:05.656 "data_size": 0 00:22:05.656 }, 00:22:05.656 { 00:22:05.656 "name": "BaseBdev3", 00:22:05.656 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:05.656 "is_configured": false, 00:22:05.656 "data_offset": 0, 00:22:05.656 "data_size": 0 00:22:05.656 } 00:22:05.656 ] 00:22:05.656 }' 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:05.656 20:21:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:05.914 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:22:05.914 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:05.914 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:05.914 [2024-10-01 20:21:01.152096] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:22:05.914 [2024-10-01 20:21:01.152153] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:22:05.915 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:05.915 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:22:05.915 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:05.915 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:05.915 [2024-10-01 20:21:01.160086] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:22:05.915 [2024-10-01 20:21:01.160140] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:22:05.915 [2024-10-01 20:21:01.160156] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:22:05.915 [2024-10-01 20:21:01.160173] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:22:05.915 [2024-10-01 20:21:01.160183] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:22:05.915 [2024-10-01 20:21:01.160198] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:22:05.915 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:05.915 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:22:05.915 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:05.915 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:06.173 [2024-10-01 20:21:01.205106] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:22:06.173 BaseBdev1 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:06.173 [ 00:22:06.173 { 00:22:06.173 "name": "BaseBdev1", 00:22:06.173 "aliases": [ 00:22:06.173 "98b9bc32-122f-47f9-bced-f7ce0a943ff0" 00:22:06.173 ], 00:22:06.173 "product_name": "Malloc disk", 00:22:06.173 "block_size": 512, 00:22:06.173 "num_blocks": 65536, 00:22:06.173 "uuid": "98b9bc32-122f-47f9-bced-f7ce0a943ff0", 00:22:06.173 "assigned_rate_limits": { 00:22:06.173 "rw_ios_per_sec": 0, 00:22:06.173 "rw_mbytes_per_sec": 0, 00:22:06.173 "r_mbytes_per_sec": 0, 00:22:06.173 "w_mbytes_per_sec": 0 00:22:06.173 }, 00:22:06.173 "claimed": true, 00:22:06.173 "claim_type": "exclusive_write", 00:22:06.173 "zoned": false, 00:22:06.173 "supported_io_types": { 00:22:06.173 "read": true, 00:22:06.173 "write": true, 00:22:06.173 "unmap": true, 00:22:06.173 "flush": true, 00:22:06.173 "reset": true, 00:22:06.173 "nvme_admin": false, 00:22:06.173 "nvme_io": false, 00:22:06.173 "nvme_io_md": false, 00:22:06.173 "write_zeroes": true, 00:22:06.173 "zcopy": true, 00:22:06.173 "get_zone_info": false, 00:22:06.173 "zone_management": false, 00:22:06.173 "zone_append": false, 00:22:06.173 "compare": false, 00:22:06.173 "compare_and_write": false, 00:22:06.173 "abort": true, 00:22:06.173 "seek_hole": false, 00:22:06.173 "seek_data": false, 00:22:06.173 "copy": true, 00:22:06.173 "nvme_iov_md": false 00:22:06.173 }, 00:22:06.173 "memory_domains": [ 00:22:06.173 { 00:22:06.173 "dma_device_id": "system", 00:22:06.173 "dma_device_type": 1 00:22:06.173 }, 00:22:06.173 { 00:22:06.173 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:06.173 "dma_device_type": 2 00:22:06.173 } 00:22:06.173 ], 00:22:06.173 "driver_specific": {} 00:22:06.173 } 00:22:06.173 ] 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:06.173 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:06.174 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:06.174 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:06.174 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:06.174 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:06.174 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:06.174 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:06.174 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:06.174 "name": "Existed_Raid", 00:22:06.174 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:06.174 "strip_size_kb": 64, 00:22:06.174 "state": "configuring", 00:22:06.174 "raid_level": "concat", 00:22:06.174 "superblock": false, 00:22:06.174 "num_base_bdevs": 3, 00:22:06.174 "num_base_bdevs_discovered": 1, 00:22:06.174 "num_base_bdevs_operational": 3, 00:22:06.174 "base_bdevs_list": [ 00:22:06.174 { 00:22:06.174 "name": "BaseBdev1", 00:22:06.174 "uuid": "98b9bc32-122f-47f9-bced-f7ce0a943ff0", 00:22:06.174 "is_configured": true, 00:22:06.174 "data_offset": 0, 00:22:06.174 "data_size": 65536 00:22:06.174 }, 00:22:06.174 { 00:22:06.174 "name": "BaseBdev2", 00:22:06.174 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:06.174 "is_configured": false, 00:22:06.174 "data_offset": 0, 00:22:06.174 "data_size": 0 00:22:06.174 }, 00:22:06.174 { 00:22:06.174 "name": "BaseBdev3", 00:22:06.174 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:06.174 "is_configured": false, 00:22:06.174 "data_offset": 0, 00:22:06.174 "data_size": 0 00:22:06.174 } 00:22:06.174 ] 00:22:06.174 }' 00:22:06.174 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:06.174 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:06.741 [2024-10-01 20:21:01.777356] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:22:06.741 [2024-10-01 20:21:01.777435] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:06.741 [2024-10-01 20:21:01.785340] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:22:06.741 [2024-10-01 20:21:01.787906] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:22:06.741 [2024-10-01 20:21:01.787962] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:22:06.741 [2024-10-01 20:21:01.787979] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:22:06.741 [2024-10-01 20:21:01.787995] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:06.741 "name": "Existed_Raid", 00:22:06.741 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:06.741 "strip_size_kb": 64, 00:22:06.741 "state": "configuring", 00:22:06.741 "raid_level": "concat", 00:22:06.741 "superblock": false, 00:22:06.741 "num_base_bdevs": 3, 00:22:06.741 "num_base_bdevs_discovered": 1, 00:22:06.741 "num_base_bdevs_operational": 3, 00:22:06.741 "base_bdevs_list": [ 00:22:06.741 { 00:22:06.741 "name": "BaseBdev1", 00:22:06.741 "uuid": "98b9bc32-122f-47f9-bced-f7ce0a943ff0", 00:22:06.741 "is_configured": true, 00:22:06.741 "data_offset": 0, 00:22:06.741 "data_size": 65536 00:22:06.741 }, 00:22:06.741 { 00:22:06.741 "name": "BaseBdev2", 00:22:06.741 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:06.741 "is_configured": false, 00:22:06.741 "data_offset": 0, 00:22:06.741 "data_size": 0 00:22:06.741 }, 00:22:06.741 { 00:22:06.741 "name": "BaseBdev3", 00:22:06.741 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:06.741 "is_configured": false, 00:22:06.741 "data_offset": 0, 00:22:06.741 "data_size": 0 00:22:06.741 } 00:22:06.741 ] 00:22:06.741 }' 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:06.741 20:21:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:07.309 [2024-10-01 20:21:02.323973] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:22:07.309 BaseBdev2 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:07.309 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:07.309 [ 00:22:07.309 { 00:22:07.309 "name": "BaseBdev2", 00:22:07.309 "aliases": [ 00:22:07.309 "82fe6dee-0be7-4720-9e78-9036d0a58d58" 00:22:07.309 ], 00:22:07.309 "product_name": "Malloc disk", 00:22:07.309 "block_size": 512, 00:22:07.309 "num_blocks": 65536, 00:22:07.309 "uuid": "82fe6dee-0be7-4720-9e78-9036d0a58d58", 00:22:07.309 "assigned_rate_limits": { 00:22:07.310 "rw_ios_per_sec": 0, 00:22:07.310 "rw_mbytes_per_sec": 0, 00:22:07.310 "r_mbytes_per_sec": 0, 00:22:07.310 "w_mbytes_per_sec": 0 00:22:07.310 }, 00:22:07.310 "claimed": true, 00:22:07.310 "claim_type": "exclusive_write", 00:22:07.310 "zoned": false, 00:22:07.310 "supported_io_types": { 00:22:07.310 "read": true, 00:22:07.310 "write": true, 00:22:07.310 "unmap": true, 00:22:07.310 "flush": true, 00:22:07.310 "reset": true, 00:22:07.310 "nvme_admin": false, 00:22:07.310 "nvme_io": false, 00:22:07.310 "nvme_io_md": false, 00:22:07.310 "write_zeroes": true, 00:22:07.310 "zcopy": true, 00:22:07.310 "get_zone_info": false, 00:22:07.310 "zone_management": false, 00:22:07.310 "zone_append": false, 00:22:07.310 "compare": false, 00:22:07.310 "compare_and_write": false, 00:22:07.310 "abort": true, 00:22:07.310 "seek_hole": false, 00:22:07.310 "seek_data": false, 00:22:07.310 "copy": true, 00:22:07.310 "nvme_iov_md": false 00:22:07.310 }, 00:22:07.310 "memory_domains": [ 00:22:07.310 { 00:22:07.310 "dma_device_id": "system", 00:22:07.310 "dma_device_type": 1 00:22:07.310 }, 00:22:07.310 { 00:22:07.310 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:07.310 "dma_device_type": 2 00:22:07.310 } 00:22:07.310 ], 00:22:07.310 "driver_specific": {} 00:22:07.310 } 00:22:07.310 ] 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:07.310 "name": "Existed_Raid", 00:22:07.310 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:07.310 "strip_size_kb": 64, 00:22:07.310 "state": "configuring", 00:22:07.310 "raid_level": "concat", 00:22:07.310 "superblock": false, 00:22:07.310 "num_base_bdevs": 3, 00:22:07.310 "num_base_bdevs_discovered": 2, 00:22:07.310 "num_base_bdevs_operational": 3, 00:22:07.310 "base_bdevs_list": [ 00:22:07.310 { 00:22:07.310 "name": "BaseBdev1", 00:22:07.310 "uuid": "98b9bc32-122f-47f9-bced-f7ce0a943ff0", 00:22:07.310 "is_configured": true, 00:22:07.310 "data_offset": 0, 00:22:07.310 "data_size": 65536 00:22:07.310 }, 00:22:07.310 { 00:22:07.310 "name": "BaseBdev2", 00:22:07.310 "uuid": "82fe6dee-0be7-4720-9e78-9036d0a58d58", 00:22:07.310 "is_configured": true, 00:22:07.310 "data_offset": 0, 00:22:07.310 "data_size": 65536 00:22:07.310 }, 00:22:07.310 { 00:22:07.310 "name": "BaseBdev3", 00:22:07.310 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:07.310 "is_configured": false, 00:22:07.310 "data_offset": 0, 00:22:07.310 "data_size": 0 00:22:07.310 } 00:22:07.310 ] 00:22:07.310 }' 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:07.310 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:07.881 [2024-10-01 20:21:02.882902] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:22:07.881 [2024-10-01 20:21:02.882975] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:22:07.881 [2024-10-01 20:21:02.882997] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:22:07.881 [2024-10-01 20:21:02.883340] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:22:07.881 [2024-10-01 20:21:02.883564] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:22:07.881 [2024-10-01 20:21:02.883582] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:22:07.881 [2024-10-01 20:21:02.883932] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:07.881 BaseBdev3 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:07.881 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:07.881 [ 00:22:07.881 { 00:22:07.881 "name": "BaseBdev3", 00:22:07.881 "aliases": [ 00:22:07.881 "88fd9592-56fc-4fdd-95f5-6ca67c827124" 00:22:07.881 ], 00:22:07.881 "product_name": "Malloc disk", 00:22:07.881 "block_size": 512, 00:22:07.881 "num_blocks": 65536, 00:22:07.882 "uuid": "88fd9592-56fc-4fdd-95f5-6ca67c827124", 00:22:07.882 "assigned_rate_limits": { 00:22:07.882 "rw_ios_per_sec": 0, 00:22:07.882 "rw_mbytes_per_sec": 0, 00:22:07.882 "r_mbytes_per_sec": 0, 00:22:07.882 "w_mbytes_per_sec": 0 00:22:07.882 }, 00:22:07.882 "claimed": true, 00:22:07.882 "claim_type": "exclusive_write", 00:22:07.882 "zoned": false, 00:22:07.882 "supported_io_types": { 00:22:07.882 "read": true, 00:22:07.882 "write": true, 00:22:07.882 "unmap": true, 00:22:07.882 "flush": true, 00:22:07.882 "reset": true, 00:22:07.882 "nvme_admin": false, 00:22:07.882 "nvme_io": false, 00:22:07.882 "nvme_io_md": false, 00:22:07.882 "write_zeroes": true, 00:22:07.882 "zcopy": true, 00:22:07.882 "get_zone_info": false, 00:22:07.882 "zone_management": false, 00:22:07.882 "zone_append": false, 00:22:07.882 "compare": false, 00:22:07.882 "compare_and_write": false, 00:22:07.882 "abort": true, 00:22:07.882 "seek_hole": false, 00:22:07.882 "seek_data": false, 00:22:07.882 "copy": true, 00:22:07.882 "nvme_iov_md": false 00:22:07.882 }, 00:22:07.882 "memory_domains": [ 00:22:07.882 { 00:22:07.882 "dma_device_id": "system", 00:22:07.882 "dma_device_type": 1 00:22:07.882 }, 00:22:07.882 { 00:22:07.882 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:07.882 "dma_device_type": 2 00:22:07.882 } 00:22:07.882 ], 00:22:07.882 "driver_specific": {} 00:22:07.882 } 00:22:07.882 ] 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:07.882 "name": "Existed_Raid", 00:22:07.882 "uuid": "9621dbf7-18e7-43ad-b52b-c4eb7864ce9b", 00:22:07.882 "strip_size_kb": 64, 00:22:07.882 "state": "online", 00:22:07.882 "raid_level": "concat", 00:22:07.882 "superblock": false, 00:22:07.882 "num_base_bdevs": 3, 00:22:07.882 "num_base_bdevs_discovered": 3, 00:22:07.882 "num_base_bdevs_operational": 3, 00:22:07.882 "base_bdevs_list": [ 00:22:07.882 { 00:22:07.882 "name": "BaseBdev1", 00:22:07.882 "uuid": "98b9bc32-122f-47f9-bced-f7ce0a943ff0", 00:22:07.882 "is_configured": true, 00:22:07.882 "data_offset": 0, 00:22:07.882 "data_size": 65536 00:22:07.882 }, 00:22:07.882 { 00:22:07.882 "name": "BaseBdev2", 00:22:07.882 "uuid": "82fe6dee-0be7-4720-9e78-9036d0a58d58", 00:22:07.882 "is_configured": true, 00:22:07.882 "data_offset": 0, 00:22:07.882 "data_size": 65536 00:22:07.882 }, 00:22:07.882 { 00:22:07.882 "name": "BaseBdev3", 00:22:07.882 "uuid": "88fd9592-56fc-4fdd-95f5-6ca67c827124", 00:22:07.882 "is_configured": true, 00:22:07.882 "data_offset": 0, 00:22:07.882 "data_size": 65536 00:22:07.882 } 00:22:07.882 ] 00:22:07.882 }' 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:07.882 20:21:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:08.448 [2024-10-01 20:21:03.439501] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:22:08.448 "name": "Existed_Raid", 00:22:08.448 "aliases": [ 00:22:08.448 "9621dbf7-18e7-43ad-b52b-c4eb7864ce9b" 00:22:08.448 ], 00:22:08.448 "product_name": "Raid Volume", 00:22:08.448 "block_size": 512, 00:22:08.448 "num_blocks": 196608, 00:22:08.448 "uuid": "9621dbf7-18e7-43ad-b52b-c4eb7864ce9b", 00:22:08.448 "assigned_rate_limits": { 00:22:08.448 "rw_ios_per_sec": 0, 00:22:08.448 "rw_mbytes_per_sec": 0, 00:22:08.448 "r_mbytes_per_sec": 0, 00:22:08.448 "w_mbytes_per_sec": 0 00:22:08.448 }, 00:22:08.448 "claimed": false, 00:22:08.448 "zoned": false, 00:22:08.448 "supported_io_types": { 00:22:08.448 "read": true, 00:22:08.448 "write": true, 00:22:08.448 "unmap": true, 00:22:08.448 "flush": true, 00:22:08.448 "reset": true, 00:22:08.448 "nvme_admin": false, 00:22:08.448 "nvme_io": false, 00:22:08.448 "nvme_io_md": false, 00:22:08.448 "write_zeroes": true, 00:22:08.448 "zcopy": false, 00:22:08.448 "get_zone_info": false, 00:22:08.448 "zone_management": false, 00:22:08.448 "zone_append": false, 00:22:08.448 "compare": false, 00:22:08.448 "compare_and_write": false, 00:22:08.448 "abort": false, 00:22:08.448 "seek_hole": false, 00:22:08.448 "seek_data": false, 00:22:08.448 "copy": false, 00:22:08.448 "nvme_iov_md": false 00:22:08.448 }, 00:22:08.448 "memory_domains": [ 00:22:08.448 { 00:22:08.448 "dma_device_id": "system", 00:22:08.448 "dma_device_type": 1 00:22:08.448 }, 00:22:08.448 { 00:22:08.448 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:08.448 "dma_device_type": 2 00:22:08.448 }, 00:22:08.448 { 00:22:08.448 "dma_device_id": "system", 00:22:08.448 "dma_device_type": 1 00:22:08.448 }, 00:22:08.448 { 00:22:08.448 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:08.448 "dma_device_type": 2 00:22:08.448 }, 00:22:08.448 { 00:22:08.448 "dma_device_id": "system", 00:22:08.448 "dma_device_type": 1 00:22:08.448 }, 00:22:08.448 { 00:22:08.448 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:08.448 "dma_device_type": 2 00:22:08.448 } 00:22:08.448 ], 00:22:08.448 "driver_specific": { 00:22:08.448 "raid": { 00:22:08.448 "uuid": "9621dbf7-18e7-43ad-b52b-c4eb7864ce9b", 00:22:08.448 "strip_size_kb": 64, 00:22:08.448 "state": "online", 00:22:08.448 "raid_level": "concat", 00:22:08.448 "superblock": false, 00:22:08.448 "num_base_bdevs": 3, 00:22:08.448 "num_base_bdevs_discovered": 3, 00:22:08.448 "num_base_bdevs_operational": 3, 00:22:08.448 "base_bdevs_list": [ 00:22:08.448 { 00:22:08.448 "name": "BaseBdev1", 00:22:08.448 "uuid": "98b9bc32-122f-47f9-bced-f7ce0a943ff0", 00:22:08.448 "is_configured": true, 00:22:08.448 "data_offset": 0, 00:22:08.448 "data_size": 65536 00:22:08.448 }, 00:22:08.448 { 00:22:08.448 "name": "BaseBdev2", 00:22:08.448 "uuid": "82fe6dee-0be7-4720-9e78-9036d0a58d58", 00:22:08.448 "is_configured": true, 00:22:08.448 "data_offset": 0, 00:22:08.448 "data_size": 65536 00:22:08.448 }, 00:22:08.448 { 00:22:08.448 "name": "BaseBdev3", 00:22:08.448 "uuid": "88fd9592-56fc-4fdd-95f5-6ca67c827124", 00:22:08.448 "is_configured": true, 00:22:08.448 "data_offset": 0, 00:22:08.448 "data_size": 65536 00:22:08.448 } 00:22:08.448 ] 00:22:08.448 } 00:22:08.448 } 00:22:08.448 }' 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:22:08.448 BaseBdev2 00:22:08.448 BaseBdev3' 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:08.448 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:08.449 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:08.706 [2024-10-01 20:21:03.743244] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:22:08.706 [2024-10-01 20:21:03.743286] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:22:08.706 [2024-10-01 20:21:03.743367] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:08.706 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:08.707 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:08.707 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:08.707 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:08.707 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:08.707 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:08.707 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:08.707 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:08.707 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:08.707 "name": "Existed_Raid", 00:22:08.707 "uuid": "9621dbf7-18e7-43ad-b52b-c4eb7864ce9b", 00:22:08.707 "strip_size_kb": 64, 00:22:08.707 "state": "offline", 00:22:08.707 "raid_level": "concat", 00:22:08.707 "superblock": false, 00:22:08.707 "num_base_bdevs": 3, 00:22:08.707 "num_base_bdevs_discovered": 2, 00:22:08.707 "num_base_bdevs_operational": 2, 00:22:08.707 "base_bdevs_list": [ 00:22:08.707 { 00:22:08.707 "name": null, 00:22:08.707 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:08.707 "is_configured": false, 00:22:08.707 "data_offset": 0, 00:22:08.707 "data_size": 65536 00:22:08.707 }, 00:22:08.707 { 00:22:08.707 "name": "BaseBdev2", 00:22:08.707 "uuid": "82fe6dee-0be7-4720-9e78-9036d0a58d58", 00:22:08.707 "is_configured": true, 00:22:08.707 "data_offset": 0, 00:22:08.707 "data_size": 65536 00:22:08.707 }, 00:22:08.707 { 00:22:08.707 "name": "BaseBdev3", 00:22:08.707 "uuid": "88fd9592-56fc-4fdd-95f5-6ca67c827124", 00:22:08.707 "is_configured": true, 00:22:08.707 "data_offset": 0, 00:22:08.707 "data_size": 65536 00:22:08.707 } 00:22:08.707 ] 00:22:08.707 }' 00:22:08.707 20:21:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:08.707 20:21:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:09.273 [2024-10-01 20:21:04.394809] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:22:09.273 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:09.533 [2024-10-01 20:21:04.533523] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:22:09.533 [2024-10-01 20:21:04.533604] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:09.533 BaseBdev2 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:09.533 [ 00:22:09.533 { 00:22:09.533 "name": "BaseBdev2", 00:22:09.533 "aliases": [ 00:22:09.533 "3d03d169-2eaa-4bed-b102-d9f7ecead208" 00:22:09.533 ], 00:22:09.533 "product_name": "Malloc disk", 00:22:09.533 "block_size": 512, 00:22:09.533 "num_blocks": 65536, 00:22:09.533 "uuid": "3d03d169-2eaa-4bed-b102-d9f7ecead208", 00:22:09.533 "assigned_rate_limits": { 00:22:09.533 "rw_ios_per_sec": 0, 00:22:09.533 "rw_mbytes_per_sec": 0, 00:22:09.533 "r_mbytes_per_sec": 0, 00:22:09.533 "w_mbytes_per_sec": 0 00:22:09.533 }, 00:22:09.533 "claimed": false, 00:22:09.533 "zoned": false, 00:22:09.533 "supported_io_types": { 00:22:09.533 "read": true, 00:22:09.533 "write": true, 00:22:09.533 "unmap": true, 00:22:09.533 "flush": true, 00:22:09.533 "reset": true, 00:22:09.533 "nvme_admin": false, 00:22:09.533 "nvme_io": false, 00:22:09.533 "nvme_io_md": false, 00:22:09.533 "write_zeroes": true, 00:22:09.533 "zcopy": true, 00:22:09.533 "get_zone_info": false, 00:22:09.533 "zone_management": false, 00:22:09.533 "zone_append": false, 00:22:09.533 "compare": false, 00:22:09.533 "compare_and_write": false, 00:22:09.533 "abort": true, 00:22:09.533 "seek_hole": false, 00:22:09.533 "seek_data": false, 00:22:09.533 "copy": true, 00:22:09.533 "nvme_iov_md": false 00:22:09.533 }, 00:22:09.533 "memory_domains": [ 00:22:09.533 { 00:22:09.533 "dma_device_id": "system", 00:22:09.533 "dma_device_type": 1 00:22:09.533 }, 00:22:09.533 { 00:22:09.533 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:09.533 "dma_device_type": 2 00:22:09.533 } 00:22:09.533 ], 00:22:09.533 "driver_specific": {} 00:22:09.533 } 00:22:09.533 ] 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:09.533 BaseBdev3 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:09.533 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:09.793 [ 00:22:09.793 { 00:22:09.793 "name": "BaseBdev3", 00:22:09.793 "aliases": [ 00:22:09.793 "63e73e37-a032-496e-ac0f-d11a889bff4a" 00:22:09.793 ], 00:22:09.793 "product_name": "Malloc disk", 00:22:09.793 "block_size": 512, 00:22:09.793 "num_blocks": 65536, 00:22:09.793 "uuid": "63e73e37-a032-496e-ac0f-d11a889bff4a", 00:22:09.793 "assigned_rate_limits": { 00:22:09.793 "rw_ios_per_sec": 0, 00:22:09.793 "rw_mbytes_per_sec": 0, 00:22:09.793 "r_mbytes_per_sec": 0, 00:22:09.793 "w_mbytes_per_sec": 0 00:22:09.793 }, 00:22:09.793 "claimed": false, 00:22:09.793 "zoned": false, 00:22:09.793 "supported_io_types": { 00:22:09.793 "read": true, 00:22:09.793 "write": true, 00:22:09.793 "unmap": true, 00:22:09.793 "flush": true, 00:22:09.793 "reset": true, 00:22:09.793 "nvme_admin": false, 00:22:09.793 "nvme_io": false, 00:22:09.793 "nvme_io_md": false, 00:22:09.793 "write_zeroes": true, 00:22:09.793 "zcopy": true, 00:22:09.793 "get_zone_info": false, 00:22:09.793 "zone_management": false, 00:22:09.793 "zone_append": false, 00:22:09.793 "compare": false, 00:22:09.793 "compare_and_write": false, 00:22:09.793 "abort": true, 00:22:09.793 "seek_hole": false, 00:22:09.793 "seek_data": false, 00:22:09.793 "copy": true, 00:22:09.793 "nvme_iov_md": false 00:22:09.793 }, 00:22:09.793 "memory_domains": [ 00:22:09.793 { 00:22:09.793 "dma_device_id": "system", 00:22:09.793 "dma_device_type": 1 00:22:09.793 }, 00:22:09.793 { 00:22:09.793 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:09.793 "dma_device_type": 2 00:22:09.793 } 00:22:09.793 ], 00:22:09.793 "driver_specific": {} 00:22:09.793 } 00:22:09.793 ] 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:09.793 [2024-10-01 20:21:04.807296] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:22:09.793 [2024-10-01 20:21:04.807355] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:22:09.793 [2024-10-01 20:21:04.807387] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:22:09.793 [2024-10-01 20:21:04.809942] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:09.793 "name": "Existed_Raid", 00:22:09.793 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:09.793 "strip_size_kb": 64, 00:22:09.793 "state": "configuring", 00:22:09.793 "raid_level": "concat", 00:22:09.793 "superblock": false, 00:22:09.793 "num_base_bdevs": 3, 00:22:09.793 "num_base_bdevs_discovered": 2, 00:22:09.793 "num_base_bdevs_operational": 3, 00:22:09.793 "base_bdevs_list": [ 00:22:09.793 { 00:22:09.793 "name": "BaseBdev1", 00:22:09.793 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:09.793 "is_configured": false, 00:22:09.793 "data_offset": 0, 00:22:09.793 "data_size": 0 00:22:09.793 }, 00:22:09.793 { 00:22:09.793 "name": "BaseBdev2", 00:22:09.793 "uuid": "3d03d169-2eaa-4bed-b102-d9f7ecead208", 00:22:09.793 "is_configured": true, 00:22:09.793 "data_offset": 0, 00:22:09.793 "data_size": 65536 00:22:09.793 }, 00:22:09.793 { 00:22:09.793 "name": "BaseBdev3", 00:22:09.793 "uuid": "63e73e37-a032-496e-ac0f-d11a889bff4a", 00:22:09.793 "is_configured": true, 00:22:09.793 "data_offset": 0, 00:22:09.793 "data_size": 65536 00:22:09.793 } 00:22:09.793 ] 00:22:09.793 }' 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:09.793 20:21:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:10.051 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:22:10.051 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:10.051 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:10.051 [2024-10-01 20:21:05.291421] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:22:10.051 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:10.051 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:10.051 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:10.051 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:10.051 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:10.051 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:10.051 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:10.051 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:10.051 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:10.051 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:10.051 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:10.051 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:10.051 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:10.052 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:10.052 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:10.310 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:10.310 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:10.310 "name": "Existed_Raid", 00:22:10.310 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:10.310 "strip_size_kb": 64, 00:22:10.310 "state": "configuring", 00:22:10.310 "raid_level": "concat", 00:22:10.310 "superblock": false, 00:22:10.310 "num_base_bdevs": 3, 00:22:10.310 "num_base_bdevs_discovered": 1, 00:22:10.310 "num_base_bdevs_operational": 3, 00:22:10.310 "base_bdevs_list": [ 00:22:10.310 { 00:22:10.310 "name": "BaseBdev1", 00:22:10.310 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:10.310 "is_configured": false, 00:22:10.310 "data_offset": 0, 00:22:10.310 "data_size": 0 00:22:10.310 }, 00:22:10.310 { 00:22:10.310 "name": null, 00:22:10.310 "uuid": "3d03d169-2eaa-4bed-b102-d9f7ecead208", 00:22:10.310 "is_configured": false, 00:22:10.310 "data_offset": 0, 00:22:10.310 "data_size": 65536 00:22:10.310 }, 00:22:10.310 { 00:22:10.310 "name": "BaseBdev3", 00:22:10.310 "uuid": "63e73e37-a032-496e-ac0f-d11a889bff4a", 00:22:10.310 "is_configured": true, 00:22:10.310 "data_offset": 0, 00:22:10.310 "data_size": 65536 00:22:10.310 } 00:22:10.310 ] 00:22:10.310 }' 00:22:10.310 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:10.310 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:10.568 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:10.568 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:22:10.568 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:10.568 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:10.568 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:10.826 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:22:10.826 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:22:10.826 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:10.826 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:10.826 [2024-10-01 20:21:05.866269] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:22:10.826 BaseBdev1 00:22:10.826 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:10.826 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:22:10.826 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:22:10.826 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:10.826 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:10.827 [ 00:22:10.827 { 00:22:10.827 "name": "BaseBdev1", 00:22:10.827 "aliases": [ 00:22:10.827 "e7b2b1f2-94bf-4311-8121-a3d144075972" 00:22:10.827 ], 00:22:10.827 "product_name": "Malloc disk", 00:22:10.827 "block_size": 512, 00:22:10.827 "num_blocks": 65536, 00:22:10.827 "uuid": "e7b2b1f2-94bf-4311-8121-a3d144075972", 00:22:10.827 "assigned_rate_limits": { 00:22:10.827 "rw_ios_per_sec": 0, 00:22:10.827 "rw_mbytes_per_sec": 0, 00:22:10.827 "r_mbytes_per_sec": 0, 00:22:10.827 "w_mbytes_per_sec": 0 00:22:10.827 }, 00:22:10.827 "claimed": true, 00:22:10.827 "claim_type": "exclusive_write", 00:22:10.827 "zoned": false, 00:22:10.827 "supported_io_types": { 00:22:10.827 "read": true, 00:22:10.827 "write": true, 00:22:10.827 "unmap": true, 00:22:10.827 "flush": true, 00:22:10.827 "reset": true, 00:22:10.827 "nvme_admin": false, 00:22:10.827 "nvme_io": false, 00:22:10.827 "nvme_io_md": false, 00:22:10.827 "write_zeroes": true, 00:22:10.827 "zcopy": true, 00:22:10.827 "get_zone_info": false, 00:22:10.827 "zone_management": false, 00:22:10.827 "zone_append": false, 00:22:10.827 "compare": false, 00:22:10.827 "compare_and_write": false, 00:22:10.827 "abort": true, 00:22:10.827 "seek_hole": false, 00:22:10.827 "seek_data": false, 00:22:10.827 "copy": true, 00:22:10.827 "nvme_iov_md": false 00:22:10.827 }, 00:22:10.827 "memory_domains": [ 00:22:10.827 { 00:22:10.827 "dma_device_id": "system", 00:22:10.827 "dma_device_type": 1 00:22:10.827 }, 00:22:10.827 { 00:22:10.827 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:10.827 "dma_device_type": 2 00:22:10.827 } 00:22:10.827 ], 00:22:10.827 "driver_specific": {} 00:22:10.827 } 00:22:10.827 ] 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:10.827 "name": "Existed_Raid", 00:22:10.827 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:10.827 "strip_size_kb": 64, 00:22:10.827 "state": "configuring", 00:22:10.827 "raid_level": "concat", 00:22:10.827 "superblock": false, 00:22:10.827 "num_base_bdevs": 3, 00:22:10.827 "num_base_bdevs_discovered": 2, 00:22:10.827 "num_base_bdevs_operational": 3, 00:22:10.827 "base_bdevs_list": [ 00:22:10.827 { 00:22:10.827 "name": "BaseBdev1", 00:22:10.827 "uuid": "e7b2b1f2-94bf-4311-8121-a3d144075972", 00:22:10.827 "is_configured": true, 00:22:10.827 "data_offset": 0, 00:22:10.827 "data_size": 65536 00:22:10.827 }, 00:22:10.827 { 00:22:10.827 "name": null, 00:22:10.827 "uuid": "3d03d169-2eaa-4bed-b102-d9f7ecead208", 00:22:10.827 "is_configured": false, 00:22:10.827 "data_offset": 0, 00:22:10.827 "data_size": 65536 00:22:10.827 }, 00:22:10.827 { 00:22:10.827 "name": "BaseBdev3", 00:22:10.827 "uuid": "63e73e37-a032-496e-ac0f-d11a889bff4a", 00:22:10.827 "is_configured": true, 00:22:10.827 "data_offset": 0, 00:22:10.827 "data_size": 65536 00:22:10.827 } 00:22:10.827 ] 00:22:10.827 }' 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:10.827 20:21:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:11.423 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:11.423 20:21:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:11.423 20:21:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:11.423 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:22:11.423 20:21:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:11.423 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:22:11.423 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:22:11.423 20:21:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:11.423 20:21:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:11.423 [2024-10-01 20:21:06.466543] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:22:11.423 20:21:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:11.423 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:11.423 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:11.423 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:11.423 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:11.424 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:11.424 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:11.424 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:11.424 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:11.424 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:11.424 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:11.424 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:11.424 20:21:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:11.424 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:11.424 20:21:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:11.424 20:21:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:11.424 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:11.424 "name": "Existed_Raid", 00:22:11.424 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:11.424 "strip_size_kb": 64, 00:22:11.424 "state": "configuring", 00:22:11.424 "raid_level": "concat", 00:22:11.424 "superblock": false, 00:22:11.424 "num_base_bdevs": 3, 00:22:11.424 "num_base_bdevs_discovered": 1, 00:22:11.424 "num_base_bdevs_operational": 3, 00:22:11.424 "base_bdevs_list": [ 00:22:11.424 { 00:22:11.424 "name": "BaseBdev1", 00:22:11.424 "uuid": "e7b2b1f2-94bf-4311-8121-a3d144075972", 00:22:11.424 "is_configured": true, 00:22:11.424 "data_offset": 0, 00:22:11.424 "data_size": 65536 00:22:11.424 }, 00:22:11.424 { 00:22:11.424 "name": null, 00:22:11.424 "uuid": "3d03d169-2eaa-4bed-b102-d9f7ecead208", 00:22:11.424 "is_configured": false, 00:22:11.424 "data_offset": 0, 00:22:11.424 "data_size": 65536 00:22:11.424 }, 00:22:11.424 { 00:22:11.424 "name": null, 00:22:11.424 "uuid": "63e73e37-a032-496e-ac0f-d11a889bff4a", 00:22:11.424 "is_configured": false, 00:22:11.424 "data_offset": 0, 00:22:11.424 "data_size": 65536 00:22:11.424 } 00:22:11.424 ] 00:22:11.424 }' 00:22:11.424 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:11.424 20:21:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:11.990 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:11.990 20:21:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:22:11.990 20:21:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:11.990 20:21:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:11.990 20:21:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:11.990 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:11.991 [2024-10-01 20:21:07.030651] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:11.991 "name": "Existed_Raid", 00:22:11.991 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:11.991 "strip_size_kb": 64, 00:22:11.991 "state": "configuring", 00:22:11.991 "raid_level": "concat", 00:22:11.991 "superblock": false, 00:22:11.991 "num_base_bdevs": 3, 00:22:11.991 "num_base_bdevs_discovered": 2, 00:22:11.991 "num_base_bdevs_operational": 3, 00:22:11.991 "base_bdevs_list": [ 00:22:11.991 { 00:22:11.991 "name": "BaseBdev1", 00:22:11.991 "uuid": "e7b2b1f2-94bf-4311-8121-a3d144075972", 00:22:11.991 "is_configured": true, 00:22:11.991 "data_offset": 0, 00:22:11.991 "data_size": 65536 00:22:11.991 }, 00:22:11.991 { 00:22:11.991 "name": null, 00:22:11.991 "uuid": "3d03d169-2eaa-4bed-b102-d9f7ecead208", 00:22:11.991 "is_configured": false, 00:22:11.991 "data_offset": 0, 00:22:11.991 "data_size": 65536 00:22:11.991 }, 00:22:11.991 { 00:22:11.991 "name": "BaseBdev3", 00:22:11.991 "uuid": "63e73e37-a032-496e-ac0f-d11a889bff4a", 00:22:11.991 "is_configured": true, 00:22:11.991 "data_offset": 0, 00:22:11.991 "data_size": 65536 00:22:11.991 } 00:22:11.991 ] 00:22:11.991 }' 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:11.991 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:12.558 [2024-10-01 20:21:07.638922] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:12.558 "name": "Existed_Raid", 00:22:12.558 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:12.558 "strip_size_kb": 64, 00:22:12.558 "state": "configuring", 00:22:12.558 "raid_level": "concat", 00:22:12.558 "superblock": false, 00:22:12.558 "num_base_bdevs": 3, 00:22:12.558 "num_base_bdevs_discovered": 1, 00:22:12.558 "num_base_bdevs_operational": 3, 00:22:12.558 "base_bdevs_list": [ 00:22:12.558 { 00:22:12.558 "name": null, 00:22:12.558 "uuid": "e7b2b1f2-94bf-4311-8121-a3d144075972", 00:22:12.558 "is_configured": false, 00:22:12.558 "data_offset": 0, 00:22:12.558 "data_size": 65536 00:22:12.558 }, 00:22:12.558 { 00:22:12.558 "name": null, 00:22:12.558 "uuid": "3d03d169-2eaa-4bed-b102-d9f7ecead208", 00:22:12.558 "is_configured": false, 00:22:12.558 "data_offset": 0, 00:22:12.558 "data_size": 65536 00:22:12.558 }, 00:22:12.558 { 00:22:12.558 "name": "BaseBdev3", 00:22:12.558 "uuid": "63e73e37-a032-496e-ac0f-d11a889bff4a", 00:22:12.558 "is_configured": true, 00:22:12.558 "data_offset": 0, 00:22:12.558 "data_size": 65536 00:22:12.558 } 00:22:12.558 ] 00:22:12.558 }' 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:12.558 20:21:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:13.126 [2024-10-01 20:21:08.269131] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:13.126 "name": "Existed_Raid", 00:22:13.126 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:13.126 "strip_size_kb": 64, 00:22:13.126 "state": "configuring", 00:22:13.126 "raid_level": "concat", 00:22:13.126 "superblock": false, 00:22:13.126 "num_base_bdevs": 3, 00:22:13.126 "num_base_bdevs_discovered": 2, 00:22:13.126 "num_base_bdevs_operational": 3, 00:22:13.126 "base_bdevs_list": [ 00:22:13.126 { 00:22:13.126 "name": null, 00:22:13.126 "uuid": "e7b2b1f2-94bf-4311-8121-a3d144075972", 00:22:13.126 "is_configured": false, 00:22:13.126 "data_offset": 0, 00:22:13.126 "data_size": 65536 00:22:13.126 }, 00:22:13.126 { 00:22:13.126 "name": "BaseBdev2", 00:22:13.126 "uuid": "3d03d169-2eaa-4bed-b102-d9f7ecead208", 00:22:13.126 "is_configured": true, 00:22:13.126 "data_offset": 0, 00:22:13.126 "data_size": 65536 00:22:13.126 }, 00:22:13.126 { 00:22:13.126 "name": "BaseBdev3", 00:22:13.126 "uuid": "63e73e37-a032-496e-ac0f-d11a889bff4a", 00:22:13.126 "is_configured": true, 00:22:13.126 "data_offset": 0, 00:22:13.126 "data_size": 65536 00:22:13.126 } 00:22:13.126 ] 00:22:13.126 }' 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:13.126 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u e7b2b1f2-94bf-4311-8121-a3d144075972 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:13.695 [2024-10-01 20:21:08.935706] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:22:13.695 [2024-10-01 20:21:08.935788] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:22:13.695 [2024-10-01 20:21:08.935805] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:22:13.695 [2024-10-01 20:21:08.936154] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:22:13.695 [2024-10-01 20:21:08.936344] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:22:13.695 [2024-10-01 20:21:08.936360] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:22:13.695 [2024-10-01 20:21:08.936675] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:13.695 NewBaseBdev 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:13.695 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:13.955 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:13.955 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:22:13.955 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:13.955 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:13.955 [ 00:22:13.955 { 00:22:13.955 "name": "NewBaseBdev", 00:22:13.955 "aliases": [ 00:22:13.955 "e7b2b1f2-94bf-4311-8121-a3d144075972" 00:22:13.955 ], 00:22:13.955 "product_name": "Malloc disk", 00:22:13.955 "block_size": 512, 00:22:13.955 "num_blocks": 65536, 00:22:13.955 "uuid": "e7b2b1f2-94bf-4311-8121-a3d144075972", 00:22:13.955 "assigned_rate_limits": { 00:22:13.955 "rw_ios_per_sec": 0, 00:22:13.955 "rw_mbytes_per_sec": 0, 00:22:13.955 "r_mbytes_per_sec": 0, 00:22:13.955 "w_mbytes_per_sec": 0 00:22:13.955 }, 00:22:13.955 "claimed": true, 00:22:13.955 "claim_type": "exclusive_write", 00:22:13.955 "zoned": false, 00:22:13.955 "supported_io_types": { 00:22:13.955 "read": true, 00:22:13.955 "write": true, 00:22:13.955 "unmap": true, 00:22:13.955 "flush": true, 00:22:13.955 "reset": true, 00:22:13.955 "nvme_admin": false, 00:22:13.955 "nvme_io": false, 00:22:13.955 "nvme_io_md": false, 00:22:13.955 "write_zeroes": true, 00:22:13.955 "zcopy": true, 00:22:13.955 "get_zone_info": false, 00:22:13.955 "zone_management": false, 00:22:13.955 "zone_append": false, 00:22:13.955 "compare": false, 00:22:13.955 "compare_and_write": false, 00:22:13.955 "abort": true, 00:22:13.955 "seek_hole": false, 00:22:13.955 "seek_data": false, 00:22:13.955 "copy": true, 00:22:13.955 "nvme_iov_md": false 00:22:13.955 }, 00:22:13.955 "memory_domains": [ 00:22:13.955 { 00:22:13.955 "dma_device_id": "system", 00:22:13.955 "dma_device_type": 1 00:22:13.955 }, 00:22:13.955 { 00:22:13.955 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:13.955 "dma_device_type": 2 00:22:13.955 } 00:22:13.955 ], 00:22:13.955 "driver_specific": {} 00:22:13.955 } 00:22:13.955 ] 00:22:13.955 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:13.955 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:22:13.955 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:22:13.955 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:13.955 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:22:13.955 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:13.955 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:13.955 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:13.955 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:13.956 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:13.956 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:13.956 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:13.956 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:13.956 20:21:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:13.956 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:13.956 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:13.956 20:21:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:13.956 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:13.956 "name": "Existed_Raid", 00:22:13.956 "uuid": "4aeb257a-1619-48da-8ed0-c5acb0f2b41f", 00:22:13.956 "strip_size_kb": 64, 00:22:13.956 "state": "online", 00:22:13.956 "raid_level": "concat", 00:22:13.956 "superblock": false, 00:22:13.956 "num_base_bdevs": 3, 00:22:13.956 "num_base_bdevs_discovered": 3, 00:22:13.956 "num_base_bdevs_operational": 3, 00:22:13.956 "base_bdevs_list": [ 00:22:13.956 { 00:22:13.956 "name": "NewBaseBdev", 00:22:13.956 "uuid": "e7b2b1f2-94bf-4311-8121-a3d144075972", 00:22:13.956 "is_configured": true, 00:22:13.956 "data_offset": 0, 00:22:13.956 "data_size": 65536 00:22:13.956 }, 00:22:13.956 { 00:22:13.956 "name": "BaseBdev2", 00:22:13.956 "uuid": "3d03d169-2eaa-4bed-b102-d9f7ecead208", 00:22:13.956 "is_configured": true, 00:22:13.956 "data_offset": 0, 00:22:13.956 "data_size": 65536 00:22:13.956 }, 00:22:13.956 { 00:22:13.956 "name": "BaseBdev3", 00:22:13.956 "uuid": "63e73e37-a032-496e-ac0f-d11a889bff4a", 00:22:13.956 "is_configured": true, 00:22:13.956 "data_offset": 0, 00:22:13.956 "data_size": 65536 00:22:13.956 } 00:22:13.956 ] 00:22:13.956 }' 00:22:13.956 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:13.956 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:14.524 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:22:14.524 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:22:14.524 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:22:14.524 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:22:14.524 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:22:14.524 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:22:14.524 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:22:14.524 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:22:14.524 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:14.524 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:14.524 [2024-10-01 20:21:09.500330] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:14.524 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:14.524 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:22:14.524 "name": "Existed_Raid", 00:22:14.524 "aliases": [ 00:22:14.524 "4aeb257a-1619-48da-8ed0-c5acb0f2b41f" 00:22:14.524 ], 00:22:14.524 "product_name": "Raid Volume", 00:22:14.524 "block_size": 512, 00:22:14.524 "num_blocks": 196608, 00:22:14.524 "uuid": "4aeb257a-1619-48da-8ed0-c5acb0f2b41f", 00:22:14.524 "assigned_rate_limits": { 00:22:14.524 "rw_ios_per_sec": 0, 00:22:14.524 "rw_mbytes_per_sec": 0, 00:22:14.524 "r_mbytes_per_sec": 0, 00:22:14.524 "w_mbytes_per_sec": 0 00:22:14.524 }, 00:22:14.524 "claimed": false, 00:22:14.524 "zoned": false, 00:22:14.524 "supported_io_types": { 00:22:14.524 "read": true, 00:22:14.524 "write": true, 00:22:14.524 "unmap": true, 00:22:14.524 "flush": true, 00:22:14.524 "reset": true, 00:22:14.524 "nvme_admin": false, 00:22:14.524 "nvme_io": false, 00:22:14.524 "nvme_io_md": false, 00:22:14.524 "write_zeroes": true, 00:22:14.524 "zcopy": false, 00:22:14.524 "get_zone_info": false, 00:22:14.524 "zone_management": false, 00:22:14.524 "zone_append": false, 00:22:14.524 "compare": false, 00:22:14.524 "compare_and_write": false, 00:22:14.524 "abort": false, 00:22:14.524 "seek_hole": false, 00:22:14.524 "seek_data": false, 00:22:14.524 "copy": false, 00:22:14.524 "nvme_iov_md": false 00:22:14.524 }, 00:22:14.524 "memory_domains": [ 00:22:14.524 { 00:22:14.524 "dma_device_id": "system", 00:22:14.524 "dma_device_type": 1 00:22:14.524 }, 00:22:14.524 { 00:22:14.524 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:14.524 "dma_device_type": 2 00:22:14.524 }, 00:22:14.524 { 00:22:14.524 "dma_device_id": "system", 00:22:14.524 "dma_device_type": 1 00:22:14.524 }, 00:22:14.524 { 00:22:14.524 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:14.524 "dma_device_type": 2 00:22:14.524 }, 00:22:14.524 { 00:22:14.525 "dma_device_id": "system", 00:22:14.525 "dma_device_type": 1 00:22:14.525 }, 00:22:14.525 { 00:22:14.525 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:14.525 "dma_device_type": 2 00:22:14.525 } 00:22:14.525 ], 00:22:14.525 "driver_specific": { 00:22:14.525 "raid": { 00:22:14.525 "uuid": "4aeb257a-1619-48da-8ed0-c5acb0f2b41f", 00:22:14.525 "strip_size_kb": 64, 00:22:14.525 "state": "online", 00:22:14.525 "raid_level": "concat", 00:22:14.525 "superblock": false, 00:22:14.525 "num_base_bdevs": 3, 00:22:14.525 "num_base_bdevs_discovered": 3, 00:22:14.525 "num_base_bdevs_operational": 3, 00:22:14.525 "base_bdevs_list": [ 00:22:14.525 { 00:22:14.525 "name": "NewBaseBdev", 00:22:14.525 "uuid": "e7b2b1f2-94bf-4311-8121-a3d144075972", 00:22:14.525 "is_configured": true, 00:22:14.525 "data_offset": 0, 00:22:14.525 "data_size": 65536 00:22:14.525 }, 00:22:14.525 { 00:22:14.525 "name": "BaseBdev2", 00:22:14.525 "uuid": "3d03d169-2eaa-4bed-b102-d9f7ecead208", 00:22:14.525 "is_configured": true, 00:22:14.525 "data_offset": 0, 00:22:14.525 "data_size": 65536 00:22:14.525 }, 00:22:14.525 { 00:22:14.525 "name": "BaseBdev3", 00:22:14.525 "uuid": "63e73e37-a032-496e-ac0f-d11a889bff4a", 00:22:14.525 "is_configured": true, 00:22:14.525 "data_offset": 0, 00:22:14.525 "data_size": 65536 00:22:14.525 } 00:22:14.525 ] 00:22:14.525 } 00:22:14.525 } 00:22:14.525 }' 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:22:14.525 BaseBdev2 00:22:14.525 BaseBdev3' 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:14.525 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:14.784 [2024-10-01 20:21:09.828023] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:22:14.784 [2024-10-01 20:21:09.828066] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:22:14.784 [2024-10-01 20:21:09.828185] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:14.784 [2024-10-01 20:21:09.828267] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:22:14.784 [2024-10-01 20:21:09.828288] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 66174 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 66174 ']' 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 66174 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 66174 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:22:14.784 killing process with pid 66174 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 66174' 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 66174 00:22:14.784 [2024-10-01 20:21:09.868122] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:22:14.784 20:21:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 66174 00:22:15.042 [2024-10-01 20:21:10.141738] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:22:16.958 20:21:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:22:16.958 00:22:16.958 real 0m12.475s 00:22:16.958 user 0m20.161s 00:22:16.958 sys 0m1.699s 00:22:16.958 20:21:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:22:16.958 20:21:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:16.958 ************************************ 00:22:16.958 END TEST raid_state_function_test 00:22:16.958 ************************************ 00:22:16.958 20:21:11 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 3 true 00:22:16.958 20:21:11 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:22:16.958 20:21:11 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:22:16.958 20:21:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:22:16.958 ************************************ 00:22:16.958 START TEST raid_state_function_test_sb 00:22:16.958 ************************************ 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 3 true 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=66817 00:22:16.958 Process raid pid: 66817 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 66817' 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 66817 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 66817 ']' 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:16.958 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:16.958 20:21:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:16.958 [2024-10-01 20:21:12.147518] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:22:16.958 [2024-10-01 20:21:12.147703] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:22:17.217 [2024-10-01 20:21:12.325796] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:17.476 [2024-10-01 20:21:12.583660] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:22:17.735 [2024-10-01 20:21:12.796784] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:17.735 [2024-10-01 20:21:12.796844] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:18.304 [2024-10-01 20:21:13.298780] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:22:18.304 [2024-10-01 20:21:13.298877] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:22:18.304 [2024-10-01 20:21:13.298898] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:22:18.304 [2024-10-01 20:21:13.298918] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:22:18.304 [2024-10-01 20:21:13.298930] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:22:18.304 [2024-10-01 20:21:13.298945] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:18.304 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:18.304 "name": "Existed_Raid", 00:22:18.304 "uuid": "dc2ea6b1-79b8-46f3-9b7f-f6005b60cfc7", 00:22:18.304 "strip_size_kb": 64, 00:22:18.304 "state": "configuring", 00:22:18.304 "raid_level": "concat", 00:22:18.304 "superblock": true, 00:22:18.304 "num_base_bdevs": 3, 00:22:18.304 "num_base_bdevs_discovered": 0, 00:22:18.304 "num_base_bdevs_operational": 3, 00:22:18.304 "base_bdevs_list": [ 00:22:18.304 { 00:22:18.304 "name": "BaseBdev1", 00:22:18.304 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:18.304 "is_configured": false, 00:22:18.304 "data_offset": 0, 00:22:18.304 "data_size": 0 00:22:18.304 }, 00:22:18.304 { 00:22:18.304 "name": "BaseBdev2", 00:22:18.304 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:18.304 "is_configured": false, 00:22:18.304 "data_offset": 0, 00:22:18.304 "data_size": 0 00:22:18.304 }, 00:22:18.304 { 00:22:18.305 "name": "BaseBdev3", 00:22:18.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:18.305 "is_configured": false, 00:22:18.305 "data_offset": 0, 00:22:18.305 "data_size": 0 00:22:18.305 } 00:22:18.305 ] 00:22:18.305 }' 00:22:18.305 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:18.305 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:18.563 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:22:18.563 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:18.563 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:18.563 [2024-10-01 20:21:13.794874] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:22:18.563 [2024-10-01 20:21:13.794930] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:22:18.563 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:18.563 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:22:18.564 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:18.564 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:18.564 [2024-10-01 20:21:13.806871] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:22:18.564 [2024-10-01 20:21:13.806942] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:22:18.564 [2024-10-01 20:21:13.806957] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:22:18.564 [2024-10-01 20:21:13.806975] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:22:18.564 [2024-10-01 20:21:13.806986] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:22:18.564 [2024-10-01 20:21:13.807002] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:22:18.564 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:18.564 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:22:18.564 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:18.564 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:18.823 [2024-10-01 20:21:13.852390] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:22:18.823 BaseBdev1 00:22:18.823 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:18.823 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:22:18.823 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:22:18.823 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:18.823 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:22:18.823 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:18.823 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:18.823 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:18.823 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:18.823 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:18.823 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:18.823 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:22:18.823 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:18.823 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:18.823 [ 00:22:18.823 { 00:22:18.823 "name": "BaseBdev1", 00:22:18.823 "aliases": [ 00:22:18.823 "83e19535-c99a-4536-a0e9-7e377fc71ca5" 00:22:18.823 ], 00:22:18.823 "product_name": "Malloc disk", 00:22:18.823 "block_size": 512, 00:22:18.823 "num_blocks": 65536, 00:22:18.823 "uuid": "83e19535-c99a-4536-a0e9-7e377fc71ca5", 00:22:18.823 "assigned_rate_limits": { 00:22:18.823 "rw_ios_per_sec": 0, 00:22:18.823 "rw_mbytes_per_sec": 0, 00:22:18.823 "r_mbytes_per_sec": 0, 00:22:18.823 "w_mbytes_per_sec": 0 00:22:18.823 }, 00:22:18.823 "claimed": true, 00:22:18.823 "claim_type": "exclusive_write", 00:22:18.823 "zoned": false, 00:22:18.823 "supported_io_types": { 00:22:18.823 "read": true, 00:22:18.823 "write": true, 00:22:18.823 "unmap": true, 00:22:18.823 "flush": true, 00:22:18.823 "reset": true, 00:22:18.823 "nvme_admin": false, 00:22:18.823 "nvme_io": false, 00:22:18.823 "nvme_io_md": false, 00:22:18.823 "write_zeroes": true, 00:22:18.823 "zcopy": true, 00:22:18.823 "get_zone_info": false, 00:22:18.823 "zone_management": false, 00:22:18.823 "zone_append": false, 00:22:18.823 "compare": false, 00:22:18.823 "compare_and_write": false, 00:22:18.823 "abort": true, 00:22:18.823 "seek_hole": false, 00:22:18.823 "seek_data": false, 00:22:18.823 "copy": true, 00:22:18.823 "nvme_iov_md": false 00:22:18.823 }, 00:22:18.823 "memory_domains": [ 00:22:18.823 { 00:22:18.823 "dma_device_id": "system", 00:22:18.823 "dma_device_type": 1 00:22:18.823 }, 00:22:18.823 { 00:22:18.823 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:18.823 "dma_device_type": 2 00:22:18.823 } 00:22:18.823 ], 00:22:18.823 "driver_specific": {} 00:22:18.823 } 00:22:18.823 ] 00:22:18.823 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:18.823 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:18.824 "name": "Existed_Raid", 00:22:18.824 "uuid": "9b23e46b-d9ed-4d73-825a-b4b78cfeb6a0", 00:22:18.824 "strip_size_kb": 64, 00:22:18.824 "state": "configuring", 00:22:18.824 "raid_level": "concat", 00:22:18.824 "superblock": true, 00:22:18.824 "num_base_bdevs": 3, 00:22:18.824 "num_base_bdevs_discovered": 1, 00:22:18.824 "num_base_bdevs_operational": 3, 00:22:18.824 "base_bdevs_list": [ 00:22:18.824 { 00:22:18.824 "name": "BaseBdev1", 00:22:18.824 "uuid": "83e19535-c99a-4536-a0e9-7e377fc71ca5", 00:22:18.824 "is_configured": true, 00:22:18.824 "data_offset": 2048, 00:22:18.824 "data_size": 63488 00:22:18.824 }, 00:22:18.824 { 00:22:18.824 "name": "BaseBdev2", 00:22:18.824 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:18.824 "is_configured": false, 00:22:18.824 "data_offset": 0, 00:22:18.824 "data_size": 0 00:22:18.824 }, 00:22:18.824 { 00:22:18.824 "name": "BaseBdev3", 00:22:18.824 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:18.824 "is_configured": false, 00:22:18.824 "data_offset": 0, 00:22:18.824 "data_size": 0 00:22:18.824 } 00:22:18.824 ] 00:22:18.824 }' 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:18.824 20:21:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:19.391 [2024-10-01 20:21:14.404621] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:22:19.391 [2024-10-01 20:21:14.404866] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:19.391 [2024-10-01 20:21:14.412651] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:22:19.391 [2024-10-01 20:21:14.415436] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:22:19.391 [2024-10-01 20:21:14.415646] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:22:19.391 [2024-10-01 20:21:14.415800] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:22:19.391 [2024-10-01 20:21:14.415934] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:19.391 "name": "Existed_Raid", 00:22:19.391 "uuid": "edf9c736-7b9e-4cba-9428-f7d3af2236ab", 00:22:19.391 "strip_size_kb": 64, 00:22:19.391 "state": "configuring", 00:22:19.391 "raid_level": "concat", 00:22:19.391 "superblock": true, 00:22:19.391 "num_base_bdevs": 3, 00:22:19.391 "num_base_bdevs_discovered": 1, 00:22:19.391 "num_base_bdevs_operational": 3, 00:22:19.391 "base_bdevs_list": [ 00:22:19.391 { 00:22:19.391 "name": "BaseBdev1", 00:22:19.391 "uuid": "83e19535-c99a-4536-a0e9-7e377fc71ca5", 00:22:19.391 "is_configured": true, 00:22:19.391 "data_offset": 2048, 00:22:19.391 "data_size": 63488 00:22:19.391 }, 00:22:19.391 { 00:22:19.391 "name": "BaseBdev2", 00:22:19.391 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:19.391 "is_configured": false, 00:22:19.391 "data_offset": 0, 00:22:19.391 "data_size": 0 00:22:19.391 }, 00:22:19.391 { 00:22:19.391 "name": "BaseBdev3", 00:22:19.391 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:19.391 "is_configured": false, 00:22:19.391 "data_offset": 0, 00:22:19.391 "data_size": 0 00:22:19.391 } 00:22:19.391 ] 00:22:19.391 }' 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:19.391 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:19.960 [2024-10-01 20:21:14.980598] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:22:19.960 BaseBdev2 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:19.960 20:21:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:19.960 [ 00:22:19.960 { 00:22:19.960 "name": "BaseBdev2", 00:22:19.960 "aliases": [ 00:22:19.960 "641b0aea-6838-4f9e-ba39-9b6b4572ef95" 00:22:19.960 ], 00:22:19.960 "product_name": "Malloc disk", 00:22:19.960 "block_size": 512, 00:22:19.960 "num_blocks": 65536, 00:22:19.960 "uuid": "641b0aea-6838-4f9e-ba39-9b6b4572ef95", 00:22:19.960 "assigned_rate_limits": { 00:22:19.960 "rw_ios_per_sec": 0, 00:22:19.960 "rw_mbytes_per_sec": 0, 00:22:19.960 "r_mbytes_per_sec": 0, 00:22:19.960 "w_mbytes_per_sec": 0 00:22:19.960 }, 00:22:19.960 "claimed": true, 00:22:19.960 "claim_type": "exclusive_write", 00:22:19.960 "zoned": false, 00:22:19.960 "supported_io_types": { 00:22:19.960 "read": true, 00:22:19.960 "write": true, 00:22:19.960 "unmap": true, 00:22:19.960 "flush": true, 00:22:19.960 "reset": true, 00:22:19.960 "nvme_admin": false, 00:22:19.960 "nvme_io": false, 00:22:19.960 "nvme_io_md": false, 00:22:19.960 "write_zeroes": true, 00:22:19.960 "zcopy": true, 00:22:19.960 "get_zone_info": false, 00:22:19.960 "zone_management": false, 00:22:19.960 "zone_append": false, 00:22:19.960 "compare": false, 00:22:19.960 "compare_and_write": false, 00:22:19.960 "abort": true, 00:22:19.960 "seek_hole": false, 00:22:19.960 "seek_data": false, 00:22:19.960 "copy": true, 00:22:19.960 "nvme_iov_md": false 00:22:19.960 }, 00:22:19.960 "memory_domains": [ 00:22:19.960 { 00:22:19.960 "dma_device_id": "system", 00:22:19.960 "dma_device_type": 1 00:22:19.960 }, 00:22:19.960 { 00:22:19.960 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:19.960 "dma_device_type": 2 00:22:19.960 } 00:22:19.960 ], 00:22:19.961 "driver_specific": {} 00:22:19.961 } 00:22:19.961 ] 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:19.961 "name": "Existed_Raid", 00:22:19.961 "uuid": "edf9c736-7b9e-4cba-9428-f7d3af2236ab", 00:22:19.961 "strip_size_kb": 64, 00:22:19.961 "state": "configuring", 00:22:19.961 "raid_level": "concat", 00:22:19.961 "superblock": true, 00:22:19.961 "num_base_bdevs": 3, 00:22:19.961 "num_base_bdevs_discovered": 2, 00:22:19.961 "num_base_bdevs_operational": 3, 00:22:19.961 "base_bdevs_list": [ 00:22:19.961 { 00:22:19.961 "name": "BaseBdev1", 00:22:19.961 "uuid": "83e19535-c99a-4536-a0e9-7e377fc71ca5", 00:22:19.961 "is_configured": true, 00:22:19.961 "data_offset": 2048, 00:22:19.961 "data_size": 63488 00:22:19.961 }, 00:22:19.961 { 00:22:19.961 "name": "BaseBdev2", 00:22:19.961 "uuid": "641b0aea-6838-4f9e-ba39-9b6b4572ef95", 00:22:19.961 "is_configured": true, 00:22:19.961 "data_offset": 2048, 00:22:19.961 "data_size": 63488 00:22:19.961 }, 00:22:19.961 { 00:22:19.961 "name": "BaseBdev3", 00:22:19.961 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:19.961 "is_configured": false, 00:22:19.961 "data_offset": 0, 00:22:19.961 "data_size": 0 00:22:19.961 } 00:22:19.961 ] 00:22:19.961 }' 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:19.961 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:20.529 [2024-10-01 20:21:15.567004] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:22:20.529 [2024-10-01 20:21:15.567389] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:22:20.529 [2024-10-01 20:21:15.567448] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:22:20.529 [2024-10-01 20:21:15.567833] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:22:20.529 BaseBdev3 00:22:20.529 [2024-10-01 20:21:15.568039] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:22:20.529 [2024-10-01 20:21:15.568056] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:22:20.529 [2024-10-01 20:21:15.568292] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:20.529 [ 00:22:20.529 { 00:22:20.529 "name": "BaseBdev3", 00:22:20.529 "aliases": [ 00:22:20.529 "61ad841e-8688-4dc4-a439-1c8fdb166e96" 00:22:20.529 ], 00:22:20.529 "product_name": "Malloc disk", 00:22:20.529 "block_size": 512, 00:22:20.529 "num_blocks": 65536, 00:22:20.529 "uuid": "61ad841e-8688-4dc4-a439-1c8fdb166e96", 00:22:20.529 "assigned_rate_limits": { 00:22:20.529 "rw_ios_per_sec": 0, 00:22:20.529 "rw_mbytes_per_sec": 0, 00:22:20.529 "r_mbytes_per_sec": 0, 00:22:20.529 "w_mbytes_per_sec": 0 00:22:20.529 }, 00:22:20.529 "claimed": true, 00:22:20.529 "claim_type": "exclusive_write", 00:22:20.529 "zoned": false, 00:22:20.529 "supported_io_types": { 00:22:20.529 "read": true, 00:22:20.529 "write": true, 00:22:20.529 "unmap": true, 00:22:20.529 "flush": true, 00:22:20.529 "reset": true, 00:22:20.529 "nvme_admin": false, 00:22:20.529 "nvme_io": false, 00:22:20.529 "nvme_io_md": false, 00:22:20.529 "write_zeroes": true, 00:22:20.529 "zcopy": true, 00:22:20.529 "get_zone_info": false, 00:22:20.529 "zone_management": false, 00:22:20.529 "zone_append": false, 00:22:20.529 "compare": false, 00:22:20.529 "compare_and_write": false, 00:22:20.529 "abort": true, 00:22:20.529 "seek_hole": false, 00:22:20.529 "seek_data": false, 00:22:20.529 "copy": true, 00:22:20.529 "nvme_iov_md": false 00:22:20.529 }, 00:22:20.529 "memory_domains": [ 00:22:20.529 { 00:22:20.529 "dma_device_id": "system", 00:22:20.529 "dma_device_type": 1 00:22:20.529 }, 00:22:20.529 { 00:22:20.529 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:20.529 "dma_device_type": 2 00:22:20.529 } 00:22:20.529 ], 00:22:20.529 "driver_specific": {} 00:22:20.529 } 00:22:20.529 ] 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:20.529 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:20.529 "name": "Existed_Raid", 00:22:20.529 "uuid": "edf9c736-7b9e-4cba-9428-f7d3af2236ab", 00:22:20.529 "strip_size_kb": 64, 00:22:20.529 "state": "online", 00:22:20.529 "raid_level": "concat", 00:22:20.529 "superblock": true, 00:22:20.529 "num_base_bdevs": 3, 00:22:20.529 "num_base_bdevs_discovered": 3, 00:22:20.529 "num_base_bdevs_operational": 3, 00:22:20.529 "base_bdevs_list": [ 00:22:20.529 { 00:22:20.529 "name": "BaseBdev1", 00:22:20.530 "uuid": "83e19535-c99a-4536-a0e9-7e377fc71ca5", 00:22:20.530 "is_configured": true, 00:22:20.530 "data_offset": 2048, 00:22:20.530 "data_size": 63488 00:22:20.530 }, 00:22:20.530 { 00:22:20.530 "name": "BaseBdev2", 00:22:20.530 "uuid": "641b0aea-6838-4f9e-ba39-9b6b4572ef95", 00:22:20.530 "is_configured": true, 00:22:20.530 "data_offset": 2048, 00:22:20.530 "data_size": 63488 00:22:20.530 }, 00:22:20.530 { 00:22:20.530 "name": "BaseBdev3", 00:22:20.530 "uuid": "61ad841e-8688-4dc4-a439-1c8fdb166e96", 00:22:20.530 "is_configured": true, 00:22:20.530 "data_offset": 2048, 00:22:20.530 "data_size": 63488 00:22:20.530 } 00:22:20.530 ] 00:22:20.530 }' 00:22:20.530 20:21:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:20.530 20:21:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:21.096 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:21.097 [2024-10-01 20:21:16.151689] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:22:21.097 "name": "Existed_Raid", 00:22:21.097 "aliases": [ 00:22:21.097 "edf9c736-7b9e-4cba-9428-f7d3af2236ab" 00:22:21.097 ], 00:22:21.097 "product_name": "Raid Volume", 00:22:21.097 "block_size": 512, 00:22:21.097 "num_blocks": 190464, 00:22:21.097 "uuid": "edf9c736-7b9e-4cba-9428-f7d3af2236ab", 00:22:21.097 "assigned_rate_limits": { 00:22:21.097 "rw_ios_per_sec": 0, 00:22:21.097 "rw_mbytes_per_sec": 0, 00:22:21.097 "r_mbytes_per_sec": 0, 00:22:21.097 "w_mbytes_per_sec": 0 00:22:21.097 }, 00:22:21.097 "claimed": false, 00:22:21.097 "zoned": false, 00:22:21.097 "supported_io_types": { 00:22:21.097 "read": true, 00:22:21.097 "write": true, 00:22:21.097 "unmap": true, 00:22:21.097 "flush": true, 00:22:21.097 "reset": true, 00:22:21.097 "nvme_admin": false, 00:22:21.097 "nvme_io": false, 00:22:21.097 "nvme_io_md": false, 00:22:21.097 "write_zeroes": true, 00:22:21.097 "zcopy": false, 00:22:21.097 "get_zone_info": false, 00:22:21.097 "zone_management": false, 00:22:21.097 "zone_append": false, 00:22:21.097 "compare": false, 00:22:21.097 "compare_and_write": false, 00:22:21.097 "abort": false, 00:22:21.097 "seek_hole": false, 00:22:21.097 "seek_data": false, 00:22:21.097 "copy": false, 00:22:21.097 "nvme_iov_md": false 00:22:21.097 }, 00:22:21.097 "memory_domains": [ 00:22:21.097 { 00:22:21.097 "dma_device_id": "system", 00:22:21.097 "dma_device_type": 1 00:22:21.097 }, 00:22:21.097 { 00:22:21.097 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:21.097 "dma_device_type": 2 00:22:21.097 }, 00:22:21.097 { 00:22:21.097 "dma_device_id": "system", 00:22:21.097 "dma_device_type": 1 00:22:21.097 }, 00:22:21.097 { 00:22:21.097 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:21.097 "dma_device_type": 2 00:22:21.097 }, 00:22:21.097 { 00:22:21.097 "dma_device_id": "system", 00:22:21.097 "dma_device_type": 1 00:22:21.097 }, 00:22:21.097 { 00:22:21.097 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:21.097 "dma_device_type": 2 00:22:21.097 } 00:22:21.097 ], 00:22:21.097 "driver_specific": { 00:22:21.097 "raid": { 00:22:21.097 "uuid": "edf9c736-7b9e-4cba-9428-f7d3af2236ab", 00:22:21.097 "strip_size_kb": 64, 00:22:21.097 "state": "online", 00:22:21.097 "raid_level": "concat", 00:22:21.097 "superblock": true, 00:22:21.097 "num_base_bdevs": 3, 00:22:21.097 "num_base_bdevs_discovered": 3, 00:22:21.097 "num_base_bdevs_operational": 3, 00:22:21.097 "base_bdevs_list": [ 00:22:21.097 { 00:22:21.097 "name": "BaseBdev1", 00:22:21.097 "uuid": "83e19535-c99a-4536-a0e9-7e377fc71ca5", 00:22:21.097 "is_configured": true, 00:22:21.097 "data_offset": 2048, 00:22:21.097 "data_size": 63488 00:22:21.097 }, 00:22:21.097 { 00:22:21.097 "name": "BaseBdev2", 00:22:21.097 "uuid": "641b0aea-6838-4f9e-ba39-9b6b4572ef95", 00:22:21.097 "is_configured": true, 00:22:21.097 "data_offset": 2048, 00:22:21.097 "data_size": 63488 00:22:21.097 }, 00:22:21.097 { 00:22:21.097 "name": "BaseBdev3", 00:22:21.097 "uuid": "61ad841e-8688-4dc4-a439-1c8fdb166e96", 00:22:21.097 "is_configured": true, 00:22:21.097 "data_offset": 2048, 00:22:21.097 "data_size": 63488 00:22:21.097 } 00:22:21.097 ] 00:22:21.097 } 00:22:21.097 } 00:22:21.097 }' 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:22:21.097 BaseBdev2 00:22:21.097 BaseBdev3' 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:21.097 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:21.356 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:21.356 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:21.356 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:21.356 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:22:21.356 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:21.356 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:21.356 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:21.356 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:21.356 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:21.356 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:21.356 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:21.356 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:21.356 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:22:21.356 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:21.356 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:21.356 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:21.357 [2024-10-01 20:21:16.475435] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:22:21.357 [2024-10-01 20:21:16.475601] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:22:21.357 [2024-10-01 20:21:16.475840] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:21.357 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:21.616 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:21.616 "name": "Existed_Raid", 00:22:21.616 "uuid": "edf9c736-7b9e-4cba-9428-f7d3af2236ab", 00:22:21.616 "strip_size_kb": 64, 00:22:21.616 "state": "offline", 00:22:21.616 "raid_level": "concat", 00:22:21.616 "superblock": true, 00:22:21.616 "num_base_bdevs": 3, 00:22:21.616 "num_base_bdevs_discovered": 2, 00:22:21.616 "num_base_bdevs_operational": 2, 00:22:21.616 "base_bdevs_list": [ 00:22:21.616 { 00:22:21.616 "name": null, 00:22:21.616 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:21.616 "is_configured": false, 00:22:21.616 "data_offset": 0, 00:22:21.616 "data_size": 63488 00:22:21.616 }, 00:22:21.616 { 00:22:21.616 "name": "BaseBdev2", 00:22:21.616 "uuid": "641b0aea-6838-4f9e-ba39-9b6b4572ef95", 00:22:21.616 "is_configured": true, 00:22:21.616 "data_offset": 2048, 00:22:21.616 "data_size": 63488 00:22:21.616 }, 00:22:21.616 { 00:22:21.616 "name": "BaseBdev3", 00:22:21.616 "uuid": "61ad841e-8688-4dc4-a439-1c8fdb166e96", 00:22:21.616 "is_configured": true, 00:22:21.616 "data_offset": 2048, 00:22:21.616 "data_size": 63488 00:22:21.616 } 00:22:21.616 ] 00:22:21.616 }' 00:22:21.616 20:21:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:21.616 20:21:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:21.940 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:22:21.940 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:22:21.940 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:21.940 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:22:21.940 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:21.940 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:21.940 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:21.940 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:22:21.940 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:22:21.940 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:22:21.940 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:21.940 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:21.940 [2024-10-01 20:21:17.141521] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:22.217 [2024-10-01 20:21:17.288291] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:22:22.217 [2024-10-01 20:21:17.288492] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.217 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:22.476 BaseBdev2 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:22.476 [ 00:22:22.476 { 00:22:22.476 "name": "BaseBdev2", 00:22:22.476 "aliases": [ 00:22:22.476 "49a08073-2c0c-4fe2-acc7-3483652a7742" 00:22:22.476 ], 00:22:22.476 "product_name": "Malloc disk", 00:22:22.476 "block_size": 512, 00:22:22.476 "num_blocks": 65536, 00:22:22.476 "uuid": "49a08073-2c0c-4fe2-acc7-3483652a7742", 00:22:22.476 "assigned_rate_limits": { 00:22:22.476 "rw_ios_per_sec": 0, 00:22:22.476 "rw_mbytes_per_sec": 0, 00:22:22.476 "r_mbytes_per_sec": 0, 00:22:22.476 "w_mbytes_per_sec": 0 00:22:22.476 }, 00:22:22.476 "claimed": false, 00:22:22.476 "zoned": false, 00:22:22.476 "supported_io_types": { 00:22:22.476 "read": true, 00:22:22.476 "write": true, 00:22:22.476 "unmap": true, 00:22:22.476 "flush": true, 00:22:22.476 "reset": true, 00:22:22.476 "nvme_admin": false, 00:22:22.476 "nvme_io": false, 00:22:22.476 "nvme_io_md": false, 00:22:22.476 "write_zeroes": true, 00:22:22.476 "zcopy": true, 00:22:22.476 "get_zone_info": false, 00:22:22.476 "zone_management": false, 00:22:22.476 "zone_append": false, 00:22:22.476 "compare": false, 00:22:22.476 "compare_and_write": false, 00:22:22.476 "abort": true, 00:22:22.476 "seek_hole": false, 00:22:22.476 "seek_data": false, 00:22:22.476 "copy": true, 00:22:22.476 "nvme_iov_md": false 00:22:22.476 }, 00:22:22.476 "memory_domains": [ 00:22:22.476 { 00:22:22.476 "dma_device_id": "system", 00:22:22.476 "dma_device_type": 1 00:22:22.476 }, 00:22:22.476 { 00:22:22.476 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:22.476 "dma_device_type": 2 00:22:22.476 } 00:22:22.476 ], 00:22:22.476 "driver_specific": {} 00:22:22.476 } 00:22:22.476 ] 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:22.476 BaseBdev3 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:22.476 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:22.477 [ 00:22:22.477 { 00:22:22.477 "name": "BaseBdev3", 00:22:22.477 "aliases": [ 00:22:22.477 "181baa8e-30c1-4b4b-93a5-cfc4ec7b8ddb" 00:22:22.477 ], 00:22:22.477 "product_name": "Malloc disk", 00:22:22.477 "block_size": 512, 00:22:22.477 "num_blocks": 65536, 00:22:22.477 "uuid": "181baa8e-30c1-4b4b-93a5-cfc4ec7b8ddb", 00:22:22.477 "assigned_rate_limits": { 00:22:22.477 "rw_ios_per_sec": 0, 00:22:22.477 "rw_mbytes_per_sec": 0, 00:22:22.477 "r_mbytes_per_sec": 0, 00:22:22.477 "w_mbytes_per_sec": 0 00:22:22.477 }, 00:22:22.477 "claimed": false, 00:22:22.477 "zoned": false, 00:22:22.477 "supported_io_types": { 00:22:22.477 "read": true, 00:22:22.477 "write": true, 00:22:22.477 "unmap": true, 00:22:22.477 "flush": true, 00:22:22.477 "reset": true, 00:22:22.477 "nvme_admin": false, 00:22:22.477 "nvme_io": false, 00:22:22.477 "nvme_io_md": false, 00:22:22.477 "write_zeroes": true, 00:22:22.477 "zcopy": true, 00:22:22.477 "get_zone_info": false, 00:22:22.477 "zone_management": false, 00:22:22.477 "zone_append": false, 00:22:22.477 "compare": false, 00:22:22.477 "compare_and_write": false, 00:22:22.477 "abort": true, 00:22:22.477 "seek_hole": false, 00:22:22.477 "seek_data": false, 00:22:22.477 "copy": true, 00:22:22.477 "nvme_iov_md": false 00:22:22.477 }, 00:22:22.477 "memory_domains": [ 00:22:22.477 { 00:22:22.477 "dma_device_id": "system", 00:22:22.477 "dma_device_type": 1 00:22:22.477 }, 00:22:22.477 { 00:22:22.477 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:22.477 "dma_device_type": 2 00:22:22.477 } 00:22:22.477 ], 00:22:22.477 "driver_specific": {} 00:22:22.477 } 00:22:22.477 ] 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:22.477 [2024-10-01 20:21:17.583093] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:22:22.477 [2024-10-01 20:21:17.583274] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:22:22.477 [2024-10-01 20:21:17.583411] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:22:22.477 [2024-10-01 20:21:17.586031] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:22.477 "name": "Existed_Raid", 00:22:22.477 "uuid": "aa5d89ae-b665-4ffe-94ed-e523f2394f2c", 00:22:22.477 "strip_size_kb": 64, 00:22:22.477 "state": "configuring", 00:22:22.477 "raid_level": "concat", 00:22:22.477 "superblock": true, 00:22:22.477 "num_base_bdevs": 3, 00:22:22.477 "num_base_bdevs_discovered": 2, 00:22:22.477 "num_base_bdevs_operational": 3, 00:22:22.477 "base_bdevs_list": [ 00:22:22.477 { 00:22:22.477 "name": "BaseBdev1", 00:22:22.477 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:22.477 "is_configured": false, 00:22:22.477 "data_offset": 0, 00:22:22.477 "data_size": 0 00:22:22.477 }, 00:22:22.477 { 00:22:22.477 "name": "BaseBdev2", 00:22:22.477 "uuid": "49a08073-2c0c-4fe2-acc7-3483652a7742", 00:22:22.477 "is_configured": true, 00:22:22.477 "data_offset": 2048, 00:22:22.477 "data_size": 63488 00:22:22.477 }, 00:22:22.477 { 00:22:22.477 "name": "BaseBdev3", 00:22:22.477 "uuid": "181baa8e-30c1-4b4b-93a5-cfc4ec7b8ddb", 00:22:22.477 "is_configured": true, 00:22:22.477 "data_offset": 2048, 00:22:22.477 "data_size": 63488 00:22:22.477 } 00:22:22.477 ] 00:22:22.477 }' 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:22.477 20:21:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:23.047 [2024-10-01 20:21:18.099218] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:23.047 "name": "Existed_Raid", 00:22:23.047 "uuid": "aa5d89ae-b665-4ffe-94ed-e523f2394f2c", 00:22:23.047 "strip_size_kb": 64, 00:22:23.047 "state": "configuring", 00:22:23.047 "raid_level": "concat", 00:22:23.047 "superblock": true, 00:22:23.047 "num_base_bdevs": 3, 00:22:23.047 "num_base_bdevs_discovered": 1, 00:22:23.047 "num_base_bdevs_operational": 3, 00:22:23.047 "base_bdevs_list": [ 00:22:23.047 { 00:22:23.047 "name": "BaseBdev1", 00:22:23.047 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:23.047 "is_configured": false, 00:22:23.047 "data_offset": 0, 00:22:23.047 "data_size": 0 00:22:23.047 }, 00:22:23.047 { 00:22:23.047 "name": null, 00:22:23.047 "uuid": "49a08073-2c0c-4fe2-acc7-3483652a7742", 00:22:23.047 "is_configured": false, 00:22:23.047 "data_offset": 0, 00:22:23.047 "data_size": 63488 00:22:23.047 }, 00:22:23.047 { 00:22:23.047 "name": "BaseBdev3", 00:22:23.047 "uuid": "181baa8e-30c1-4b4b-93a5-cfc4ec7b8ddb", 00:22:23.047 "is_configured": true, 00:22:23.047 "data_offset": 2048, 00:22:23.047 "data_size": 63488 00:22:23.047 } 00:22:23.047 ] 00:22:23.047 }' 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:23.047 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:23.617 [2024-10-01 20:21:18.705820] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:22:23.617 BaseBdev1 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:23.617 [ 00:22:23.617 { 00:22:23.617 "name": "BaseBdev1", 00:22:23.617 "aliases": [ 00:22:23.617 "af638afa-2f9e-4db4-a7c9-ab0142d37fea" 00:22:23.617 ], 00:22:23.617 "product_name": "Malloc disk", 00:22:23.617 "block_size": 512, 00:22:23.617 "num_blocks": 65536, 00:22:23.617 "uuid": "af638afa-2f9e-4db4-a7c9-ab0142d37fea", 00:22:23.617 "assigned_rate_limits": { 00:22:23.617 "rw_ios_per_sec": 0, 00:22:23.617 "rw_mbytes_per_sec": 0, 00:22:23.617 "r_mbytes_per_sec": 0, 00:22:23.617 "w_mbytes_per_sec": 0 00:22:23.617 }, 00:22:23.617 "claimed": true, 00:22:23.617 "claim_type": "exclusive_write", 00:22:23.617 "zoned": false, 00:22:23.617 "supported_io_types": { 00:22:23.617 "read": true, 00:22:23.617 "write": true, 00:22:23.617 "unmap": true, 00:22:23.617 "flush": true, 00:22:23.617 "reset": true, 00:22:23.617 "nvme_admin": false, 00:22:23.617 "nvme_io": false, 00:22:23.617 "nvme_io_md": false, 00:22:23.617 "write_zeroes": true, 00:22:23.617 "zcopy": true, 00:22:23.617 "get_zone_info": false, 00:22:23.617 "zone_management": false, 00:22:23.617 "zone_append": false, 00:22:23.617 "compare": false, 00:22:23.617 "compare_and_write": false, 00:22:23.617 "abort": true, 00:22:23.617 "seek_hole": false, 00:22:23.617 "seek_data": false, 00:22:23.617 "copy": true, 00:22:23.617 "nvme_iov_md": false 00:22:23.617 }, 00:22:23.617 "memory_domains": [ 00:22:23.617 { 00:22:23.617 "dma_device_id": "system", 00:22:23.617 "dma_device_type": 1 00:22:23.617 }, 00:22:23.617 { 00:22:23.617 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:23.617 "dma_device_type": 2 00:22:23.617 } 00:22:23.617 ], 00:22:23.617 "driver_specific": {} 00:22:23.617 } 00:22:23.617 ] 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:23.617 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:23.617 "name": "Existed_Raid", 00:22:23.617 "uuid": "aa5d89ae-b665-4ffe-94ed-e523f2394f2c", 00:22:23.617 "strip_size_kb": 64, 00:22:23.617 "state": "configuring", 00:22:23.617 "raid_level": "concat", 00:22:23.617 "superblock": true, 00:22:23.617 "num_base_bdevs": 3, 00:22:23.617 "num_base_bdevs_discovered": 2, 00:22:23.617 "num_base_bdevs_operational": 3, 00:22:23.617 "base_bdevs_list": [ 00:22:23.617 { 00:22:23.617 "name": "BaseBdev1", 00:22:23.617 "uuid": "af638afa-2f9e-4db4-a7c9-ab0142d37fea", 00:22:23.617 "is_configured": true, 00:22:23.617 "data_offset": 2048, 00:22:23.617 "data_size": 63488 00:22:23.617 }, 00:22:23.617 { 00:22:23.617 "name": null, 00:22:23.617 "uuid": "49a08073-2c0c-4fe2-acc7-3483652a7742", 00:22:23.617 "is_configured": false, 00:22:23.617 "data_offset": 0, 00:22:23.617 "data_size": 63488 00:22:23.618 }, 00:22:23.618 { 00:22:23.618 "name": "BaseBdev3", 00:22:23.618 "uuid": "181baa8e-30c1-4b4b-93a5-cfc4ec7b8ddb", 00:22:23.618 "is_configured": true, 00:22:23.618 "data_offset": 2048, 00:22:23.618 "data_size": 63488 00:22:23.618 } 00:22:23.618 ] 00:22:23.618 }' 00:22:23.618 20:21:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:23.618 20:21:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:24.189 [2024-10-01 20:21:19.314101] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:24.189 "name": "Existed_Raid", 00:22:24.189 "uuid": "aa5d89ae-b665-4ffe-94ed-e523f2394f2c", 00:22:24.189 "strip_size_kb": 64, 00:22:24.189 "state": "configuring", 00:22:24.189 "raid_level": "concat", 00:22:24.189 "superblock": true, 00:22:24.189 "num_base_bdevs": 3, 00:22:24.189 "num_base_bdevs_discovered": 1, 00:22:24.189 "num_base_bdevs_operational": 3, 00:22:24.189 "base_bdevs_list": [ 00:22:24.189 { 00:22:24.189 "name": "BaseBdev1", 00:22:24.189 "uuid": "af638afa-2f9e-4db4-a7c9-ab0142d37fea", 00:22:24.189 "is_configured": true, 00:22:24.189 "data_offset": 2048, 00:22:24.189 "data_size": 63488 00:22:24.189 }, 00:22:24.189 { 00:22:24.189 "name": null, 00:22:24.189 "uuid": "49a08073-2c0c-4fe2-acc7-3483652a7742", 00:22:24.189 "is_configured": false, 00:22:24.189 "data_offset": 0, 00:22:24.189 "data_size": 63488 00:22:24.189 }, 00:22:24.189 { 00:22:24.189 "name": null, 00:22:24.189 "uuid": "181baa8e-30c1-4b4b-93a5-cfc4ec7b8ddb", 00:22:24.189 "is_configured": false, 00:22:24.189 "data_offset": 0, 00:22:24.189 "data_size": 63488 00:22:24.189 } 00:22:24.189 ] 00:22:24.189 }' 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:24.189 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:24.788 [2024-10-01 20:21:19.922251] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:24.788 "name": "Existed_Raid", 00:22:24.788 "uuid": "aa5d89ae-b665-4ffe-94ed-e523f2394f2c", 00:22:24.788 "strip_size_kb": 64, 00:22:24.788 "state": "configuring", 00:22:24.788 "raid_level": "concat", 00:22:24.788 "superblock": true, 00:22:24.788 "num_base_bdevs": 3, 00:22:24.788 "num_base_bdevs_discovered": 2, 00:22:24.788 "num_base_bdevs_operational": 3, 00:22:24.788 "base_bdevs_list": [ 00:22:24.788 { 00:22:24.788 "name": "BaseBdev1", 00:22:24.788 "uuid": "af638afa-2f9e-4db4-a7c9-ab0142d37fea", 00:22:24.788 "is_configured": true, 00:22:24.788 "data_offset": 2048, 00:22:24.788 "data_size": 63488 00:22:24.788 }, 00:22:24.788 { 00:22:24.788 "name": null, 00:22:24.788 "uuid": "49a08073-2c0c-4fe2-acc7-3483652a7742", 00:22:24.788 "is_configured": false, 00:22:24.788 "data_offset": 0, 00:22:24.788 "data_size": 63488 00:22:24.788 }, 00:22:24.788 { 00:22:24.788 "name": "BaseBdev3", 00:22:24.788 "uuid": "181baa8e-30c1-4b4b-93a5-cfc4ec7b8ddb", 00:22:24.788 "is_configured": true, 00:22:24.788 "data_offset": 2048, 00:22:24.788 "data_size": 63488 00:22:24.788 } 00:22:24.788 ] 00:22:24.788 }' 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:24.788 20:21:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:25.355 [2024-10-01 20:21:20.498453] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:25.355 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:25.615 20:21:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:25.615 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:25.615 "name": "Existed_Raid", 00:22:25.615 "uuid": "aa5d89ae-b665-4ffe-94ed-e523f2394f2c", 00:22:25.615 "strip_size_kb": 64, 00:22:25.615 "state": "configuring", 00:22:25.615 "raid_level": "concat", 00:22:25.615 "superblock": true, 00:22:25.615 "num_base_bdevs": 3, 00:22:25.615 "num_base_bdevs_discovered": 1, 00:22:25.615 "num_base_bdevs_operational": 3, 00:22:25.615 "base_bdevs_list": [ 00:22:25.615 { 00:22:25.615 "name": null, 00:22:25.615 "uuid": "af638afa-2f9e-4db4-a7c9-ab0142d37fea", 00:22:25.615 "is_configured": false, 00:22:25.615 "data_offset": 0, 00:22:25.615 "data_size": 63488 00:22:25.615 }, 00:22:25.615 { 00:22:25.615 "name": null, 00:22:25.615 "uuid": "49a08073-2c0c-4fe2-acc7-3483652a7742", 00:22:25.615 "is_configured": false, 00:22:25.615 "data_offset": 0, 00:22:25.615 "data_size": 63488 00:22:25.615 }, 00:22:25.615 { 00:22:25.615 "name": "BaseBdev3", 00:22:25.615 "uuid": "181baa8e-30c1-4b4b-93a5-cfc4ec7b8ddb", 00:22:25.615 "is_configured": true, 00:22:25.615 "data_offset": 2048, 00:22:25.615 "data_size": 63488 00:22:25.615 } 00:22:25.615 ] 00:22:25.615 }' 00:22:25.615 20:21:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:25.615 20:21:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:25.875 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:25.875 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:25.875 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:25.875 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:26.134 [2024-10-01 20:21:21.174019] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:26.134 "name": "Existed_Raid", 00:22:26.134 "uuid": "aa5d89ae-b665-4ffe-94ed-e523f2394f2c", 00:22:26.134 "strip_size_kb": 64, 00:22:26.134 "state": "configuring", 00:22:26.134 "raid_level": "concat", 00:22:26.134 "superblock": true, 00:22:26.134 "num_base_bdevs": 3, 00:22:26.134 "num_base_bdevs_discovered": 2, 00:22:26.134 "num_base_bdevs_operational": 3, 00:22:26.134 "base_bdevs_list": [ 00:22:26.134 { 00:22:26.134 "name": null, 00:22:26.134 "uuid": "af638afa-2f9e-4db4-a7c9-ab0142d37fea", 00:22:26.134 "is_configured": false, 00:22:26.134 "data_offset": 0, 00:22:26.134 "data_size": 63488 00:22:26.134 }, 00:22:26.134 { 00:22:26.134 "name": "BaseBdev2", 00:22:26.134 "uuid": "49a08073-2c0c-4fe2-acc7-3483652a7742", 00:22:26.134 "is_configured": true, 00:22:26.134 "data_offset": 2048, 00:22:26.134 "data_size": 63488 00:22:26.134 }, 00:22:26.134 { 00:22:26.134 "name": "BaseBdev3", 00:22:26.134 "uuid": "181baa8e-30c1-4b4b-93a5-cfc4ec7b8ddb", 00:22:26.134 "is_configured": true, 00:22:26.134 "data_offset": 2048, 00:22:26.134 "data_size": 63488 00:22:26.134 } 00:22:26.134 ] 00:22:26.134 }' 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:26.134 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u af638afa-2f9e-4db4-a7c9-ab0142d37fea 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:26.703 [2024-10-01 20:21:21.849017] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:22:26.703 [2024-10-01 20:21:21.849551] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:22:26.703 [2024-10-01 20:21:21.849583] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:22:26.703 NewBaseBdev 00:22:26.703 [2024-10-01 20:21:21.849944] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:22:26.703 [2024-10-01 20:21:21.850124] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:22:26.703 [2024-10-01 20:21:21.850139] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:22:26.703 [2024-10-01 20:21:21.850304] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:26.703 [ 00:22:26.703 { 00:22:26.703 "name": "NewBaseBdev", 00:22:26.703 "aliases": [ 00:22:26.703 "af638afa-2f9e-4db4-a7c9-ab0142d37fea" 00:22:26.703 ], 00:22:26.703 "product_name": "Malloc disk", 00:22:26.703 "block_size": 512, 00:22:26.703 "num_blocks": 65536, 00:22:26.703 "uuid": "af638afa-2f9e-4db4-a7c9-ab0142d37fea", 00:22:26.703 "assigned_rate_limits": { 00:22:26.703 "rw_ios_per_sec": 0, 00:22:26.703 "rw_mbytes_per_sec": 0, 00:22:26.703 "r_mbytes_per_sec": 0, 00:22:26.703 "w_mbytes_per_sec": 0 00:22:26.703 }, 00:22:26.703 "claimed": true, 00:22:26.703 "claim_type": "exclusive_write", 00:22:26.703 "zoned": false, 00:22:26.703 "supported_io_types": { 00:22:26.703 "read": true, 00:22:26.703 "write": true, 00:22:26.703 "unmap": true, 00:22:26.703 "flush": true, 00:22:26.703 "reset": true, 00:22:26.703 "nvme_admin": false, 00:22:26.703 "nvme_io": false, 00:22:26.703 "nvme_io_md": false, 00:22:26.703 "write_zeroes": true, 00:22:26.703 "zcopy": true, 00:22:26.703 "get_zone_info": false, 00:22:26.703 "zone_management": false, 00:22:26.703 "zone_append": false, 00:22:26.703 "compare": false, 00:22:26.703 "compare_and_write": false, 00:22:26.703 "abort": true, 00:22:26.703 "seek_hole": false, 00:22:26.703 "seek_data": false, 00:22:26.703 "copy": true, 00:22:26.703 "nvme_iov_md": false 00:22:26.703 }, 00:22:26.703 "memory_domains": [ 00:22:26.703 { 00:22:26.703 "dma_device_id": "system", 00:22:26.703 "dma_device_type": 1 00:22:26.703 }, 00:22:26.703 { 00:22:26.703 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:26.703 "dma_device_type": 2 00:22:26.703 } 00:22:26.703 ], 00:22:26.703 "driver_specific": {} 00:22:26.703 } 00:22:26.703 ] 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:26.703 "name": "Existed_Raid", 00:22:26.703 "uuid": "aa5d89ae-b665-4ffe-94ed-e523f2394f2c", 00:22:26.703 "strip_size_kb": 64, 00:22:26.703 "state": "online", 00:22:26.703 "raid_level": "concat", 00:22:26.703 "superblock": true, 00:22:26.703 "num_base_bdevs": 3, 00:22:26.703 "num_base_bdevs_discovered": 3, 00:22:26.703 "num_base_bdevs_operational": 3, 00:22:26.703 "base_bdevs_list": [ 00:22:26.703 { 00:22:26.703 "name": "NewBaseBdev", 00:22:26.703 "uuid": "af638afa-2f9e-4db4-a7c9-ab0142d37fea", 00:22:26.703 "is_configured": true, 00:22:26.703 "data_offset": 2048, 00:22:26.703 "data_size": 63488 00:22:26.703 }, 00:22:26.703 { 00:22:26.703 "name": "BaseBdev2", 00:22:26.703 "uuid": "49a08073-2c0c-4fe2-acc7-3483652a7742", 00:22:26.703 "is_configured": true, 00:22:26.703 "data_offset": 2048, 00:22:26.703 "data_size": 63488 00:22:26.703 }, 00:22:26.703 { 00:22:26.703 "name": "BaseBdev3", 00:22:26.703 "uuid": "181baa8e-30c1-4b4b-93a5-cfc4ec7b8ddb", 00:22:26.703 "is_configured": true, 00:22:26.703 "data_offset": 2048, 00:22:26.703 "data_size": 63488 00:22:26.703 } 00:22:26.703 ] 00:22:26.703 }' 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:26.703 20:21:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:27.269 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:22:27.269 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:22:27.269 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:22:27.269 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:22:27.269 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:22:27.269 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:22:27.269 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:22:27.269 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:22:27.269 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:27.269 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:27.269 [2024-10-01 20:21:22.405649] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:27.269 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:27.269 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:22:27.269 "name": "Existed_Raid", 00:22:27.269 "aliases": [ 00:22:27.269 "aa5d89ae-b665-4ffe-94ed-e523f2394f2c" 00:22:27.269 ], 00:22:27.269 "product_name": "Raid Volume", 00:22:27.269 "block_size": 512, 00:22:27.269 "num_blocks": 190464, 00:22:27.269 "uuid": "aa5d89ae-b665-4ffe-94ed-e523f2394f2c", 00:22:27.269 "assigned_rate_limits": { 00:22:27.269 "rw_ios_per_sec": 0, 00:22:27.269 "rw_mbytes_per_sec": 0, 00:22:27.269 "r_mbytes_per_sec": 0, 00:22:27.269 "w_mbytes_per_sec": 0 00:22:27.269 }, 00:22:27.269 "claimed": false, 00:22:27.269 "zoned": false, 00:22:27.269 "supported_io_types": { 00:22:27.269 "read": true, 00:22:27.269 "write": true, 00:22:27.269 "unmap": true, 00:22:27.269 "flush": true, 00:22:27.269 "reset": true, 00:22:27.269 "nvme_admin": false, 00:22:27.269 "nvme_io": false, 00:22:27.269 "nvme_io_md": false, 00:22:27.269 "write_zeroes": true, 00:22:27.269 "zcopy": false, 00:22:27.269 "get_zone_info": false, 00:22:27.269 "zone_management": false, 00:22:27.269 "zone_append": false, 00:22:27.269 "compare": false, 00:22:27.269 "compare_and_write": false, 00:22:27.269 "abort": false, 00:22:27.269 "seek_hole": false, 00:22:27.269 "seek_data": false, 00:22:27.269 "copy": false, 00:22:27.269 "nvme_iov_md": false 00:22:27.269 }, 00:22:27.269 "memory_domains": [ 00:22:27.269 { 00:22:27.269 "dma_device_id": "system", 00:22:27.269 "dma_device_type": 1 00:22:27.269 }, 00:22:27.269 { 00:22:27.269 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:27.269 "dma_device_type": 2 00:22:27.269 }, 00:22:27.269 { 00:22:27.269 "dma_device_id": "system", 00:22:27.269 "dma_device_type": 1 00:22:27.269 }, 00:22:27.269 { 00:22:27.269 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:27.269 "dma_device_type": 2 00:22:27.269 }, 00:22:27.269 { 00:22:27.269 "dma_device_id": "system", 00:22:27.269 "dma_device_type": 1 00:22:27.269 }, 00:22:27.269 { 00:22:27.269 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:27.269 "dma_device_type": 2 00:22:27.269 } 00:22:27.269 ], 00:22:27.269 "driver_specific": { 00:22:27.269 "raid": { 00:22:27.269 "uuid": "aa5d89ae-b665-4ffe-94ed-e523f2394f2c", 00:22:27.269 "strip_size_kb": 64, 00:22:27.269 "state": "online", 00:22:27.269 "raid_level": "concat", 00:22:27.269 "superblock": true, 00:22:27.269 "num_base_bdevs": 3, 00:22:27.269 "num_base_bdevs_discovered": 3, 00:22:27.269 "num_base_bdevs_operational": 3, 00:22:27.269 "base_bdevs_list": [ 00:22:27.269 { 00:22:27.269 "name": "NewBaseBdev", 00:22:27.269 "uuid": "af638afa-2f9e-4db4-a7c9-ab0142d37fea", 00:22:27.269 "is_configured": true, 00:22:27.269 "data_offset": 2048, 00:22:27.269 "data_size": 63488 00:22:27.269 }, 00:22:27.269 { 00:22:27.269 "name": "BaseBdev2", 00:22:27.269 "uuid": "49a08073-2c0c-4fe2-acc7-3483652a7742", 00:22:27.269 "is_configured": true, 00:22:27.269 "data_offset": 2048, 00:22:27.269 "data_size": 63488 00:22:27.269 }, 00:22:27.269 { 00:22:27.269 "name": "BaseBdev3", 00:22:27.269 "uuid": "181baa8e-30c1-4b4b-93a5-cfc4ec7b8ddb", 00:22:27.269 "is_configured": true, 00:22:27.269 "data_offset": 2048, 00:22:27.269 "data_size": 63488 00:22:27.269 } 00:22:27.269 ] 00:22:27.269 } 00:22:27.269 } 00:22:27.269 }' 00:22:27.269 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:22:27.269 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:22:27.269 BaseBdev2 00:22:27.269 BaseBdev3' 00:22:27.269 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:27.527 [2024-10-01 20:21:22.717323] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:22:27.527 [2024-10-01 20:21:22.717511] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:22:27.527 [2024-10-01 20:21:22.717743] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:27.527 [2024-10-01 20:21:22.717918] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:22:27.527 [2024-10-01 20:21:22.718052] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 66817 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 66817 ']' 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 66817 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 66817 00:22:27.527 killing process with pid 66817 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 66817' 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 66817 00:22:27.527 [2024-10-01 20:21:22.756803] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:22:27.527 20:21:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 66817 00:22:27.783 [2024-10-01 20:21:23.023196] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:22:29.685 20:21:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:22:29.685 00:22:29.685 real 0m12.822s 00:22:29.685 user 0m20.617s 00:22:29.685 sys 0m1.766s 00:22:29.685 20:21:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:22:29.685 ************************************ 00:22:29.685 END TEST raid_state_function_test_sb 00:22:29.685 ************************************ 00:22:29.685 20:21:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:22:29.685 20:21:24 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 3 00:22:29.685 20:21:24 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:22:29.685 20:21:24 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:22:29.685 20:21:24 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:22:29.685 ************************************ 00:22:29.685 START TEST raid_superblock_test 00:22:29.685 ************************************ 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test concat 3 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:22:29.685 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=67460 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 67460 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 67460 ']' 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:29.685 20:21:24 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:29.943 [2024-10-01 20:21:25.005931] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:22:29.943 [2024-10-01 20:21:25.006142] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid67460 ] 00:22:29.943 [2024-10-01 20:21:25.187299] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:30.509 [2024-10-01 20:21:25.522882] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:22:30.509 [2024-10-01 20:21:25.739995] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:30.509 [2024-10-01 20:21:25.740038] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:31.074 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:31.074 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:22:31.074 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:22:31.074 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:22:31.074 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:22:31.074 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:22:31.074 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:31.075 malloc1 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:31.075 [2024-10-01 20:21:26.208717] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:22:31.075 [2024-10-01 20:21:26.208962] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:31.075 [2024-10-01 20:21:26.209045] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:22:31.075 [2024-10-01 20:21:26.209309] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:31.075 [2024-10-01 20:21:26.212273] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:31.075 [2024-10-01 20:21:26.212320] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:22:31.075 pt1 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:31.075 malloc2 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:31.075 [2024-10-01 20:21:26.262410] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:22:31.075 [2024-10-01 20:21:26.262618] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:31.075 [2024-10-01 20:21:26.262670] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:22:31.075 [2024-10-01 20:21:26.262689] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:31.075 [2024-10-01 20:21:26.265662] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:31.075 [2024-10-01 20:21:26.265706] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:22:31.075 pt2 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:31.075 malloc3 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:31.075 [2024-10-01 20:21:26.320128] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:22:31.075 [2024-10-01 20:21:26.320218] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:31.075 [2024-10-01 20:21:26.320251] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:22:31.075 [2024-10-01 20:21:26.320267] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:31.075 [2024-10-01 20:21:26.323346] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:31.075 [2024-10-01 20:21:26.323387] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:22:31.075 pt3 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:31.075 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:31.334 [2024-10-01 20:21:26.328279] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:22:31.334 [2024-10-01 20:21:26.330941] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:22:31.334 [2024-10-01 20:21:26.331042] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:22:31.334 [2024-10-01 20:21:26.331289] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:22:31.334 [2024-10-01 20:21:26.331315] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:22:31.334 [2024-10-01 20:21:26.331645] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:22:31.334 [2024-10-01 20:21:26.331893] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:22:31.334 [2024-10-01 20:21:26.331912] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:22:31.334 [2024-10-01 20:21:26.332173] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:31.334 "name": "raid_bdev1", 00:22:31.334 "uuid": "8e4230a2-5e89-4c2b-93ea-cd409fa91b64", 00:22:31.334 "strip_size_kb": 64, 00:22:31.334 "state": "online", 00:22:31.334 "raid_level": "concat", 00:22:31.334 "superblock": true, 00:22:31.334 "num_base_bdevs": 3, 00:22:31.334 "num_base_bdevs_discovered": 3, 00:22:31.334 "num_base_bdevs_operational": 3, 00:22:31.334 "base_bdevs_list": [ 00:22:31.334 { 00:22:31.334 "name": "pt1", 00:22:31.334 "uuid": "00000000-0000-0000-0000-000000000001", 00:22:31.334 "is_configured": true, 00:22:31.334 "data_offset": 2048, 00:22:31.334 "data_size": 63488 00:22:31.334 }, 00:22:31.334 { 00:22:31.334 "name": "pt2", 00:22:31.334 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:31.334 "is_configured": true, 00:22:31.334 "data_offset": 2048, 00:22:31.334 "data_size": 63488 00:22:31.334 }, 00:22:31.334 { 00:22:31.334 "name": "pt3", 00:22:31.334 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:31.334 "is_configured": true, 00:22:31.334 "data_offset": 2048, 00:22:31.334 "data_size": 63488 00:22:31.334 } 00:22:31.334 ] 00:22:31.334 }' 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:31.334 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:31.592 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:22:31.592 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:22:31.592 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:22:31.592 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:22:31.592 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:22:31.592 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:22:31.593 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:22:31.593 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:31.593 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:22:31.593 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:31.593 [2024-10-01 20:21:26.832866] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:31.851 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:31.852 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:22:31.852 "name": "raid_bdev1", 00:22:31.852 "aliases": [ 00:22:31.852 "8e4230a2-5e89-4c2b-93ea-cd409fa91b64" 00:22:31.852 ], 00:22:31.852 "product_name": "Raid Volume", 00:22:31.852 "block_size": 512, 00:22:31.852 "num_blocks": 190464, 00:22:31.852 "uuid": "8e4230a2-5e89-4c2b-93ea-cd409fa91b64", 00:22:31.852 "assigned_rate_limits": { 00:22:31.852 "rw_ios_per_sec": 0, 00:22:31.852 "rw_mbytes_per_sec": 0, 00:22:31.852 "r_mbytes_per_sec": 0, 00:22:31.852 "w_mbytes_per_sec": 0 00:22:31.852 }, 00:22:31.852 "claimed": false, 00:22:31.852 "zoned": false, 00:22:31.852 "supported_io_types": { 00:22:31.852 "read": true, 00:22:31.852 "write": true, 00:22:31.852 "unmap": true, 00:22:31.852 "flush": true, 00:22:31.852 "reset": true, 00:22:31.852 "nvme_admin": false, 00:22:31.852 "nvme_io": false, 00:22:31.852 "nvme_io_md": false, 00:22:31.852 "write_zeroes": true, 00:22:31.852 "zcopy": false, 00:22:31.852 "get_zone_info": false, 00:22:31.852 "zone_management": false, 00:22:31.852 "zone_append": false, 00:22:31.852 "compare": false, 00:22:31.852 "compare_and_write": false, 00:22:31.852 "abort": false, 00:22:31.852 "seek_hole": false, 00:22:31.852 "seek_data": false, 00:22:31.852 "copy": false, 00:22:31.852 "nvme_iov_md": false 00:22:31.852 }, 00:22:31.852 "memory_domains": [ 00:22:31.852 { 00:22:31.852 "dma_device_id": "system", 00:22:31.852 "dma_device_type": 1 00:22:31.852 }, 00:22:31.852 { 00:22:31.852 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:31.852 "dma_device_type": 2 00:22:31.852 }, 00:22:31.852 { 00:22:31.852 "dma_device_id": "system", 00:22:31.852 "dma_device_type": 1 00:22:31.852 }, 00:22:31.852 { 00:22:31.852 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:31.852 "dma_device_type": 2 00:22:31.852 }, 00:22:31.852 { 00:22:31.852 "dma_device_id": "system", 00:22:31.852 "dma_device_type": 1 00:22:31.852 }, 00:22:31.852 { 00:22:31.852 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:31.852 "dma_device_type": 2 00:22:31.852 } 00:22:31.852 ], 00:22:31.852 "driver_specific": { 00:22:31.852 "raid": { 00:22:31.852 "uuid": "8e4230a2-5e89-4c2b-93ea-cd409fa91b64", 00:22:31.852 "strip_size_kb": 64, 00:22:31.852 "state": "online", 00:22:31.852 "raid_level": "concat", 00:22:31.852 "superblock": true, 00:22:31.852 "num_base_bdevs": 3, 00:22:31.852 "num_base_bdevs_discovered": 3, 00:22:31.852 "num_base_bdevs_operational": 3, 00:22:31.852 "base_bdevs_list": [ 00:22:31.852 { 00:22:31.852 "name": "pt1", 00:22:31.852 "uuid": "00000000-0000-0000-0000-000000000001", 00:22:31.852 "is_configured": true, 00:22:31.852 "data_offset": 2048, 00:22:31.852 "data_size": 63488 00:22:31.852 }, 00:22:31.852 { 00:22:31.852 "name": "pt2", 00:22:31.852 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:31.852 "is_configured": true, 00:22:31.852 "data_offset": 2048, 00:22:31.852 "data_size": 63488 00:22:31.852 }, 00:22:31.852 { 00:22:31.852 "name": "pt3", 00:22:31.852 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:31.852 "is_configured": true, 00:22:31.852 "data_offset": 2048, 00:22:31.852 "data_size": 63488 00:22:31.852 } 00:22:31.852 ] 00:22:31.852 } 00:22:31.852 } 00:22:31.852 }' 00:22:31.852 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:22:31.852 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:22:31.852 pt2 00:22:31.852 pt3' 00:22:31.852 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:31.852 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:22:31.852 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:31.852 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:22:31.852 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:31.852 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:31.852 20:21:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:31.852 20:21:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:31.852 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:31.852 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:31.852 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:31.852 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:22:31.852 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:31.852 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:31.852 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:31.852 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:31.852 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:31.852 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:31.852 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:31.852 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:31.852 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:22:31.852 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:31.852 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:31.852 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:32.111 [2024-10-01 20:21:27.136834] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=8e4230a2-5e89-4c2b-93ea-cd409fa91b64 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 8e4230a2-5e89-4c2b-93ea-cd409fa91b64 ']' 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:32.111 [2024-10-01 20:21:27.184479] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:22:32.111 [2024-10-01 20:21:27.184524] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:22:32.111 [2024-10-01 20:21:27.184619] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:32.111 [2024-10-01 20:21:27.184709] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:22:32.111 [2024-10-01 20:21:27.184761] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:22:32.111 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:32.112 [2024-10-01 20:21:27.340637] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:22:32.112 [2024-10-01 20:21:27.343278] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:22:32.112 [2024-10-01 20:21:27.343359] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:22:32.112 [2024-10-01 20:21:27.343436] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:22:32.112 [2024-10-01 20:21:27.343510] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:22:32.112 [2024-10-01 20:21:27.343546] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:22:32.112 [2024-10-01 20:21:27.343576] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:22:32.112 [2024-10-01 20:21:27.343591] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:22:32.112 request: 00:22:32.112 { 00:22:32.112 "name": "raid_bdev1", 00:22:32.112 "raid_level": "concat", 00:22:32.112 "base_bdevs": [ 00:22:32.112 "malloc1", 00:22:32.112 "malloc2", 00:22:32.112 "malloc3" 00:22:32.112 ], 00:22:32.112 "strip_size_kb": 64, 00:22:32.112 "superblock": false, 00:22:32.112 "method": "bdev_raid_create", 00:22:32.112 "req_id": 1 00:22:32.112 } 00:22:32.112 Got JSON-RPC error response 00:22:32.112 response: 00:22:32.112 { 00:22:32.112 "code": -17, 00:22:32.112 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:22:32.112 } 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:32.112 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:32.370 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:22:32.370 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:22:32.370 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:22:32.370 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:32.370 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:32.370 [2024-10-01 20:21:27.400558] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:22:32.370 [2024-10-01 20:21:27.400639] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:32.370 [2024-10-01 20:21:27.400673] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:22:32.370 [2024-10-01 20:21:27.400690] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:32.370 [2024-10-01 20:21:27.403765] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:32.370 [2024-10-01 20:21:27.403822] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:22:32.370 [2024-10-01 20:21:27.403940] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:22:32.370 [2024-10-01 20:21:27.404006] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:22:32.370 pt1 00:22:32.370 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:32.370 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:22:32.370 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:22:32.370 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:32.370 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:32.370 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:32.370 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:32.370 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:32.370 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:32.371 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:32.371 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:32.371 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:32.371 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:32.371 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:32.371 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:32.371 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:32.371 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:32.371 "name": "raid_bdev1", 00:22:32.371 "uuid": "8e4230a2-5e89-4c2b-93ea-cd409fa91b64", 00:22:32.371 "strip_size_kb": 64, 00:22:32.371 "state": "configuring", 00:22:32.371 "raid_level": "concat", 00:22:32.371 "superblock": true, 00:22:32.371 "num_base_bdevs": 3, 00:22:32.371 "num_base_bdevs_discovered": 1, 00:22:32.371 "num_base_bdevs_operational": 3, 00:22:32.371 "base_bdevs_list": [ 00:22:32.371 { 00:22:32.371 "name": "pt1", 00:22:32.371 "uuid": "00000000-0000-0000-0000-000000000001", 00:22:32.371 "is_configured": true, 00:22:32.371 "data_offset": 2048, 00:22:32.371 "data_size": 63488 00:22:32.371 }, 00:22:32.371 { 00:22:32.371 "name": null, 00:22:32.371 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:32.371 "is_configured": false, 00:22:32.371 "data_offset": 2048, 00:22:32.371 "data_size": 63488 00:22:32.371 }, 00:22:32.371 { 00:22:32.371 "name": null, 00:22:32.371 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:32.371 "is_configured": false, 00:22:32.371 "data_offset": 2048, 00:22:32.371 "data_size": 63488 00:22:32.371 } 00:22:32.371 ] 00:22:32.371 }' 00:22:32.371 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:32.371 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:32.936 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:22:32.936 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:22:32.936 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:32.936 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:32.937 [2024-10-01 20:21:27.916739] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:22:32.937 [2024-10-01 20:21:27.916862] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:32.937 [2024-10-01 20:21:27.916905] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:22:32.937 [2024-10-01 20:21:27.916923] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:32.937 [2024-10-01 20:21:27.917543] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:32.937 [2024-10-01 20:21:27.917577] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:22:32.937 [2024-10-01 20:21:27.917693] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:22:32.937 [2024-10-01 20:21:27.917758] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:22:32.937 pt2 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:32.937 [2024-10-01 20:21:27.924720] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:32.937 "name": "raid_bdev1", 00:22:32.937 "uuid": "8e4230a2-5e89-4c2b-93ea-cd409fa91b64", 00:22:32.937 "strip_size_kb": 64, 00:22:32.937 "state": "configuring", 00:22:32.937 "raid_level": "concat", 00:22:32.937 "superblock": true, 00:22:32.937 "num_base_bdevs": 3, 00:22:32.937 "num_base_bdevs_discovered": 1, 00:22:32.937 "num_base_bdevs_operational": 3, 00:22:32.937 "base_bdevs_list": [ 00:22:32.937 { 00:22:32.937 "name": "pt1", 00:22:32.937 "uuid": "00000000-0000-0000-0000-000000000001", 00:22:32.937 "is_configured": true, 00:22:32.937 "data_offset": 2048, 00:22:32.937 "data_size": 63488 00:22:32.937 }, 00:22:32.937 { 00:22:32.937 "name": null, 00:22:32.937 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:32.937 "is_configured": false, 00:22:32.937 "data_offset": 0, 00:22:32.937 "data_size": 63488 00:22:32.937 }, 00:22:32.937 { 00:22:32.937 "name": null, 00:22:32.937 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:32.937 "is_configured": false, 00:22:32.937 "data_offset": 2048, 00:22:32.937 "data_size": 63488 00:22:32.937 } 00:22:32.937 ] 00:22:32.937 }' 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:32.937 20:21:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:33.502 [2024-10-01 20:21:28.476868] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:22:33.502 [2024-10-01 20:21:28.476964] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:33.502 [2024-10-01 20:21:28.476996] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:22:33.502 [2024-10-01 20:21:28.477027] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:33.502 [2024-10-01 20:21:28.477625] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:33.502 [2024-10-01 20:21:28.477666] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:22:33.502 [2024-10-01 20:21:28.477796] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:22:33.502 [2024-10-01 20:21:28.477852] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:22:33.502 pt2 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:33.502 [2024-10-01 20:21:28.484847] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:22:33.502 [2024-10-01 20:21:28.484905] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:33.502 [2024-10-01 20:21:28.484929] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:22:33.502 [2024-10-01 20:21:28.484956] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:33.502 [2024-10-01 20:21:28.485406] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:33.502 [2024-10-01 20:21:28.485441] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:22:33.502 [2024-10-01 20:21:28.485517] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:22:33.502 [2024-10-01 20:21:28.485557] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:22:33.502 [2024-10-01 20:21:28.485728] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:22:33.502 [2024-10-01 20:21:28.485752] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:22:33.502 [2024-10-01 20:21:28.486079] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:22:33.502 [2024-10-01 20:21:28.486270] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:22:33.502 [2024-10-01 20:21:28.486293] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:22:33.502 [2024-10-01 20:21:28.486459] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:33.502 pt3 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:33.502 "name": "raid_bdev1", 00:22:33.502 "uuid": "8e4230a2-5e89-4c2b-93ea-cd409fa91b64", 00:22:33.502 "strip_size_kb": 64, 00:22:33.502 "state": "online", 00:22:33.502 "raid_level": "concat", 00:22:33.502 "superblock": true, 00:22:33.502 "num_base_bdevs": 3, 00:22:33.502 "num_base_bdevs_discovered": 3, 00:22:33.502 "num_base_bdevs_operational": 3, 00:22:33.502 "base_bdevs_list": [ 00:22:33.502 { 00:22:33.502 "name": "pt1", 00:22:33.502 "uuid": "00000000-0000-0000-0000-000000000001", 00:22:33.502 "is_configured": true, 00:22:33.502 "data_offset": 2048, 00:22:33.502 "data_size": 63488 00:22:33.502 }, 00:22:33.502 { 00:22:33.502 "name": "pt2", 00:22:33.502 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:33.502 "is_configured": true, 00:22:33.502 "data_offset": 2048, 00:22:33.502 "data_size": 63488 00:22:33.502 }, 00:22:33.502 { 00:22:33.502 "name": "pt3", 00:22:33.502 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:33.502 "is_configured": true, 00:22:33.502 "data_offset": 2048, 00:22:33.502 "data_size": 63488 00:22:33.502 } 00:22:33.502 ] 00:22:33.502 }' 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:33.502 20:21:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:22:34.066 [2024-10-01 20:21:29.053481] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:22:34.066 "name": "raid_bdev1", 00:22:34.066 "aliases": [ 00:22:34.066 "8e4230a2-5e89-4c2b-93ea-cd409fa91b64" 00:22:34.066 ], 00:22:34.066 "product_name": "Raid Volume", 00:22:34.066 "block_size": 512, 00:22:34.066 "num_blocks": 190464, 00:22:34.066 "uuid": "8e4230a2-5e89-4c2b-93ea-cd409fa91b64", 00:22:34.066 "assigned_rate_limits": { 00:22:34.066 "rw_ios_per_sec": 0, 00:22:34.066 "rw_mbytes_per_sec": 0, 00:22:34.066 "r_mbytes_per_sec": 0, 00:22:34.066 "w_mbytes_per_sec": 0 00:22:34.066 }, 00:22:34.066 "claimed": false, 00:22:34.066 "zoned": false, 00:22:34.066 "supported_io_types": { 00:22:34.066 "read": true, 00:22:34.066 "write": true, 00:22:34.066 "unmap": true, 00:22:34.066 "flush": true, 00:22:34.066 "reset": true, 00:22:34.066 "nvme_admin": false, 00:22:34.066 "nvme_io": false, 00:22:34.066 "nvme_io_md": false, 00:22:34.066 "write_zeroes": true, 00:22:34.066 "zcopy": false, 00:22:34.066 "get_zone_info": false, 00:22:34.066 "zone_management": false, 00:22:34.066 "zone_append": false, 00:22:34.066 "compare": false, 00:22:34.066 "compare_and_write": false, 00:22:34.066 "abort": false, 00:22:34.066 "seek_hole": false, 00:22:34.066 "seek_data": false, 00:22:34.066 "copy": false, 00:22:34.066 "nvme_iov_md": false 00:22:34.066 }, 00:22:34.066 "memory_domains": [ 00:22:34.066 { 00:22:34.066 "dma_device_id": "system", 00:22:34.066 "dma_device_type": 1 00:22:34.066 }, 00:22:34.066 { 00:22:34.066 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:34.066 "dma_device_type": 2 00:22:34.066 }, 00:22:34.066 { 00:22:34.066 "dma_device_id": "system", 00:22:34.066 "dma_device_type": 1 00:22:34.066 }, 00:22:34.066 { 00:22:34.066 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:34.066 "dma_device_type": 2 00:22:34.066 }, 00:22:34.066 { 00:22:34.066 "dma_device_id": "system", 00:22:34.066 "dma_device_type": 1 00:22:34.066 }, 00:22:34.066 { 00:22:34.066 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:34.066 "dma_device_type": 2 00:22:34.066 } 00:22:34.066 ], 00:22:34.066 "driver_specific": { 00:22:34.066 "raid": { 00:22:34.066 "uuid": "8e4230a2-5e89-4c2b-93ea-cd409fa91b64", 00:22:34.066 "strip_size_kb": 64, 00:22:34.066 "state": "online", 00:22:34.066 "raid_level": "concat", 00:22:34.066 "superblock": true, 00:22:34.066 "num_base_bdevs": 3, 00:22:34.066 "num_base_bdevs_discovered": 3, 00:22:34.066 "num_base_bdevs_operational": 3, 00:22:34.066 "base_bdevs_list": [ 00:22:34.066 { 00:22:34.066 "name": "pt1", 00:22:34.066 "uuid": "00000000-0000-0000-0000-000000000001", 00:22:34.066 "is_configured": true, 00:22:34.066 "data_offset": 2048, 00:22:34.066 "data_size": 63488 00:22:34.066 }, 00:22:34.066 { 00:22:34.066 "name": "pt2", 00:22:34.066 "uuid": "00000000-0000-0000-0000-000000000002", 00:22:34.066 "is_configured": true, 00:22:34.066 "data_offset": 2048, 00:22:34.066 "data_size": 63488 00:22:34.066 }, 00:22:34.066 { 00:22:34.066 "name": "pt3", 00:22:34.066 "uuid": "00000000-0000-0000-0000-000000000003", 00:22:34.066 "is_configured": true, 00:22:34.066 "data_offset": 2048, 00:22:34.066 "data_size": 63488 00:22:34.066 } 00:22:34.066 ] 00:22:34.066 } 00:22:34.066 } 00:22:34.066 }' 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:22:34.066 pt2 00:22:34.066 pt3' 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:34.066 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:34.067 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:34.067 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:34.067 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:34.067 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:22:34.067 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:34.067 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:34.067 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:34.323 [2024-10-01 20:21:29.369487] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 8e4230a2-5e89-4c2b-93ea-cd409fa91b64 '!=' 8e4230a2-5e89-4c2b-93ea-cd409fa91b64 ']' 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 67460 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 67460 ']' 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 67460 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 67460 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:22:34.323 killing process with pid 67460 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 67460' 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 67460 00:22:34.323 [2024-10-01 20:21:29.445706] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:22:34.323 20:21:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 67460 00:22:34.323 [2024-10-01 20:21:29.445856] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:34.323 [2024-10-01 20:21:29.445942] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:22:34.323 [2024-10-01 20:21:29.445967] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:22:34.580 [2024-10-01 20:21:29.719822] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:22:36.480 20:21:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:22:36.480 00:22:36.480 real 0m6.647s 00:22:36.480 user 0m9.463s 00:22:36.480 sys 0m0.982s 00:22:36.480 20:21:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:22:36.480 ************************************ 00:22:36.480 END TEST raid_superblock_test 00:22:36.480 ************************************ 00:22:36.480 20:21:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:22:36.480 20:21:31 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 3 read 00:22:36.480 20:21:31 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:22:36.480 20:21:31 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:22:36.480 20:21:31 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:22:36.480 ************************************ 00:22:36.480 START TEST raid_read_error_test 00:22:36.480 ************************************ 00:22:36.480 20:21:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 3 read 00:22:36.480 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:22:36.480 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:22:36.480 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:22:36.480 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:22:36.480 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:22:36.480 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.jrcHCoHOxk 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=67724 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 67724 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 67724 ']' 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:36.481 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:36.481 20:21:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:36.481 [2024-10-01 20:21:31.722405] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:22:36.481 [2024-10-01 20:21:31.722599] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid67724 ] 00:22:36.739 [2024-10-01 20:21:31.892762] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:36.999 [2024-10-01 20:21:32.139430] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:22:37.258 [2024-10-01 20:21:32.346802] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:37.258 [2024-10-01 20:21:32.346903] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:37.825 BaseBdev1_malloc 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:37.825 true 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:37.825 [2024-10-01 20:21:32.824595] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:22:37.825 [2024-10-01 20:21:32.824665] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:37.825 [2024-10-01 20:21:32.824698] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:22:37.825 [2024-10-01 20:21:32.824737] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:37.825 [2024-10-01 20:21:32.827686] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:37.825 [2024-10-01 20:21:32.827762] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:22:37.825 BaseBdev1 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:37.825 BaseBdev2_malloc 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:37.825 true 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:37.825 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:37.825 [2024-10-01 20:21:32.881592] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:22:37.825 [2024-10-01 20:21:32.881668] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:37.825 [2024-10-01 20:21:32.881699] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:22:37.825 [2024-10-01 20:21:32.881735] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:37.826 [2024-10-01 20:21:32.884699] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:37.826 [2024-10-01 20:21:32.884762] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:22:37.826 BaseBdev2 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:37.826 BaseBdev3_malloc 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:37.826 true 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:37.826 [2024-10-01 20:21:32.942716] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:22:37.826 [2024-10-01 20:21:32.942826] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:37.826 [2024-10-01 20:21:32.942863] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:22:37.826 [2024-10-01 20:21:32.942886] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:37.826 [2024-10-01 20:21:32.946079] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:37.826 [2024-10-01 20:21:32.946165] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:22:37.826 BaseBdev3 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:37.826 [2024-10-01 20:21:32.950987] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:22:37.826 [2024-10-01 20:21:32.953822] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:22:37.826 [2024-10-01 20:21:32.953958] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:22:37.826 [2024-10-01 20:21:32.954310] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:22:37.826 [2024-10-01 20:21:32.954331] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:22:37.826 [2024-10-01 20:21:32.954704] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:22:37.826 [2024-10-01 20:21:32.954970] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:22:37.826 [2024-10-01 20:21:32.954993] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:22:37.826 [2024-10-01 20:21:32.955277] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:37.826 20:21:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:37.826 20:21:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:37.826 "name": "raid_bdev1", 00:22:37.826 "uuid": "e36b1aa0-94ab-4417-94fd-1246c28113dd", 00:22:37.826 "strip_size_kb": 64, 00:22:37.826 "state": "online", 00:22:37.826 "raid_level": "concat", 00:22:37.826 "superblock": true, 00:22:37.826 "num_base_bdevs": 3, 00:22:37.826 "num_base_bdevs_discovered": 3, 00:22:37.826 "num_base_bdevs_operational": 3, 00:22:37.826 "base_bdevs_list": [ 00:22:37.826 { 00:22:37.826 "name": "BaseBdev1", 00:22:37.826 "uuid": "2c531789-38cb-5e48-937b-093edc8e8b3d", 00:22:37.826 "is_configured": true, 00:22:37.826 "data_offset": 2048, 00:22:37.826 "data_size": 63488 00:22:37.826 }, 00:22:37.826 { 00:22:37.826 "name": "BaseBdev2", 00:22:37.826 "uuid": "84778415-50c2-5d74-accf-4ac029015419", 00:22:37.826 "is_configured": true, 00:22:37.826 "data_offset": 2048, 00:22:37.826 "data_size": 63488 00:22:37.826 }, 00:22:37.826 { 00:22:37.826 "name": "BaseBdev3", 00:22:37.826 "uuid": "6fef635c-2426-5860-b153-570b3e7fa9eb", 00:22:37.826 "is_configured": true, 00:22:37.826 "data_offset": 2048, 00:22:37.826 "data_size": 63488 00:22:37.826 } 00:22:37.826 ] 00:22:37.826 }' 00:22:37.826 20:21:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:37.826 20:21:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:38.393 20:21:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:22:38.394 20:21:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:22:38.394 [2024-10-01 20:21:33.632989] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:39.328 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:39.328 "name": "raid_bdev1", 00:22:39.328 "uuid": "e36b1aa0-94ab-4417-94fd-1246c28113dd", 00:22:39.328 "strip_size_kb": 64, 00:22:39.328 "state": "online", 00:22:39.328 "raid_level": "concat", 00:22:39.329 "superblock": true, 00:22:39.329 "num_base_bdevs": 3, 00:22:39.329 "num_base_bdevs_discovered": 3, 00:22:39.329 "num_base_bdevs_operational": 3, 00:22:39.329 "base_bdevs_list": [ 00:22:39.329 { 00:22:39.329 "name": "BaseBdev1", 00:22:39.329 "uuid": "2c531789-38cb-5e48-937b-093edc8e8b3d", 00:22:39.329 "is_configured": true, 00:22:39.329 "data_offset": 2048, 00:22:39.329 "data_size": 63488 00:22:39.329 }, 00:22:39.329 { 00:22:39.329 "name": "BaseBdev2", 00:22:39.329 "uuid": "84778415-50c2-5d74-accf-4ac029015419", 00:22:39.329 "is_configured": true, 00:22:39.329 "data_offset": 2048, 00:22:39.329 "data_size": 63488 00:22:39.329 }, 00:22:39.329 { 00:22:39.329 "name": "BaseBdev3", 00:22:39.329 "uuid": "6fef635c-2426-5860-b153-570b3e7fa9eb", 00:22:39.329 "is_configured": true, 00:22:39.329 "data_offset": 2048, 00:22:39.329 "data_size": 63488 00:22:39.329 } 00:22:39.329 ] 00:22:39.329 }' 00:22:39.329 20:21:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:39.329 20:21:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:39.895 20:21:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:22:39.895 20:21:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:39.895 20:21:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:39.895 [2024-10-01 20:21:35.044498] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:22:39.895 [2024-10-01 20:21:35.044542] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:22:39.895 [2024-10-01 20:21:35.047926] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:39.895 [2024-10-01 20:21:35.048003] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:39.895 [2024-10-01 20:21:35.048061] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:22:39.895 [2024-10-01 20:21:35.048077] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:22:39.895 { 00:22:39.895 "results": [ 00:22:39.895 { 00:22:39.895 "job": "raid_bdev1", 00:22:39.895 "core_mask": "0x1", 00:22:39.895 "workload": "randrw", 00:22:39.895 "percentage": 50, 00:22:39.895 "status": "finished", 00:22:39.895 "queue_depth": 1, 00:22:39.895 "io_size": 131072, 00:22:39.895 "runtime": 1.408771, 00:22:39.895 "iops": 10019.371494728384, 00:22:39.895 "mibps": 1252.421436841048, 00:22:39.895 "io_failed": 1, 00:22:39.895 "io_timeout": 0, 00:22:39.895 "avg_latency_us": 139.62368530874056, 00:22:39.895 "min_latency_us": 41.89090909090909, 00:22:39.895 "max_latency_us": 1839.4763636363637 00:22:39.895 } 00:22:39.895 ], 00:22:39.895 "core_count": 1 00:22:39.895 } 00:22:39.895 20:21:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:39.895 20:21:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 67724 00:22:39.895 20:21:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 67724 ']' 00:22:39.895 20:21:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 67724 00:22:39.895 20:21:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:22:39.895 20:21:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:39.895 20:21:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 67724 00:22:39.895 20:21:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:22:39.895 killing process with pid 67724 00:22:39.895 20:21:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:22:39.895 20:21:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 67724' 00:22:39.895 20:21:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 67724 00:22:39.895 [2024-10-01 20:21:35.084002] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:22:39.895 20:21:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 67724 00:22:40.154 [2024-10-01 20:21:35.293561] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:22:42.055 20:21:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.jrcHCoHOxk 00:22:42.055 20:21:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:22:42.055 20:21:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:22:42.055 20:21:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.71 00:22:42.055 20:21:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:22:42.055 20:21:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:22:42.055 20:21:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:22:42.055 20:21:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.71 != \0\.\0\0 ]] 00:22:42.055 00:22:42.055 real 0m5.573s 00:22:42.055 user 0m6.587s 00:22:42.055 sys 0m0.727s 00:22:42.055 20:21:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:22:42.055 ************************************ 00:22:42.055 END TEST raid_read_error_test 00:22:42.055 ************************************ 00:22:42.055 20:21:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:42.055 20:21:37 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 3 write 00:22:42.055 20:21:37 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:22:42.055 20:21:37 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:22:42.055 20:21:37 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:22:42.055 ************************************ 00:22:42.055 START TEST raid_write_error_test 00:22:42.055 ************************************ 00:22:42.055 20:21:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 3 write 00:22:42.055 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:22:42.055 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:22:42.055 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:22:42.055 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:22:42.055 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:22:42.055 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:22:42.055 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:22:42.055 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:22:42.055 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:22:42.055 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:22:42.055 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:22:42.055 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:22:42.055 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:22:42.055 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.Uhwa73Zq1j 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=67885 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 67885 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 67885 ']' 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:42.056 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:42.056 20:21:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:42.313 [2024-10-01 20:21:37.351693] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:22:42.313 [2024-10-01 20:21:37.351925] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid67885 ] 00:22:42.313 [2024-10-01 20:21:37.527374] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:42.571 [2024-10-01 20:21:37.776323] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:22:42.830 [2024-10-01 20:21:37.982706] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:42.830 [2024-10-01 20:21:37.982796] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:43.398 BaseBdev1_malloc 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:43.398 true 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:43.398 [2024-10-01 20:21:38.458204] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:22:43.398 [2024-10-01 20:21:38.458310] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:43.398 [2024-10-01 20:21:38.458345] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:22:43.398 [2024-10-01 20:21:38.458367] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:43.398 [2024-10-01 20:21:38.461825] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:43.398 [2024-10-01 20:21:38.461879] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:22:43.398 BaseBdev1 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:43.398 BaseBdev2_malloc 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:43.398 true 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:43.398 [2024-10-01 20:21:38.519546] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:22:43.398 [2024-10-01 20:21:38.519619] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:43.398 [2024-10-01 20:21:38.519648] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:22:43.398 [2024-10-01 20:21:38.519667] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:43.398 [2024-10-01 20:21:38.522604] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:43.398 [2024-10-01 20:21:38.522656] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:22:43.398 BaseBdev2 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:43.398 BaseBdev3_malloc 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:43.398 true 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:43.398 [2024-10-01 20:21:38.581125] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:22:43.398 [2024-10-01 20:21:38.581196] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:22:43.398 [2024-10-01 20:21:38.581224] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:22:43.398 [2024-10-01 20:21:38.581245] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:22:43.398 [2024-10-01 20:21:38.584202] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:22:43.398 [2024-10-01 20:21:38.584267] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:22:43.398 BaseBdev3 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:43.398 [2024-10-01 20:21:38.589263] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:22:43.398 [2024-10-01 20:21:38.591799] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:22:43.398 [2024-10-01 20:21:38.591939] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:22:43.398 [2024-10-01 20:21:38.592228] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:22:43.398 [2024-10-01 20:21:38.592248] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:22:43.398 [2024-10-01 20:21:38.592581] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:22:43.398 [2024-10-01 20:21:38.592815] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:22:43.398 [2024-10-01 20:21:38.592838] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:22:43.398 [2024-10-01 20:21:38.593057] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:43.398 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:43.398 "name": "raid_bdev1", 00:22:43.398 "uuid": "0d7be41d-ac7c-43a5-9d64-4def2c0e0eb3", 00:22:43.398 "strip_size_kb": 64, 00:22:43.398 "state": "online", 00:22:43.398 "raid_level": "concat", 00:22:43.398 "superblock": true, 00:22:43.398 "num_base_bdevs": 3, 00:22:43.398 "num_base_bdevs_discovered": 3, 00:22:43.398 "num_base_bdevs_operational": 3, 00:22:43.398 "base_bdevs_list": [ 00:22:43.398 { 00:22:43.398 "name": "BaseBdev1", 00:22:43.398 "uuid": "da101723-8dc5-5df1-8b6a-808a96bf2281", 00:22:43.398 "is_configured": true, 00:22:43.398 "data_offset": 2048, 00:22:43.398 "data_size": 63488 00:22:43.398 }, 00:22:43.398 { 00:22:43.398 "name": "BaseBdev2", 00:22:43.398 "uuid": "49a2fc71-16bf-53eb-b3d7-4477702cd667", 00:22:43.398 "is_configured": true, 00:22:43.398 "data_offset": 2048, 00:22:43.399 "data_size": 63488 00:22:43.399 }, 00:22:43.399 { 00:22:43.399 "name": "BaseBdev3", 00:22:43.399 "uuid": "c6ec75d3-d9a0-53b5-bcdb-6b9953600be0", 00:22:43.399 "is_configured": true, 00:22:43.399 "data_offset": 2048, 00:22:43.399 "data_size": 63488 00:22:43.399 } 00:22:43.399 ] 00:22:43.399 }' 00:22:43.399 20:21:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:43.399 20:21:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:44.023 20:21:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:22:44.023 20:21:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:22:44.023 [2024-10-01 20:21:39.242842] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:22:44.956 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:22:44.956 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:44.956 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:44.956 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:44.956 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:22:44.956 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:22:44.956 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:44.957 "name": "raid_bdev1", 00:22:44.957 "uuid": "0d7be41d-ac7c-43a5-9d64-4def2c0e0eb3", 00:22:44.957 "strip_size_kb": 64, 00:22:44.957 "state": "online", 00:22:44.957 "raid_level": "concat", 00:22:44.957 "superblock": true, 00:22:44.957 "num_base_bdevs": 3, 00:22:44.957 "num_base_bdevs_discovered": 3, 00:22:44.957 "num_base_bdevs_operational": 3, 00:22:44.957 "base_bdevs_list": [ 00:22:44.957 { 00:22:44.957 "name": "BaseBdev1", 00:22:44.957 "uuid": "da101723-8dc5-5df1-8b6a-808a96bf2281", 00:22:44.957 "is_configured": true, 00:22:44.957 "data_offset": 2048, 00:22:44.957 "data_size": 63488 00:22:44.957 }, 00:22:44.957 { 00:22:44.957 "name": "BaseBdev2", 00:22:44.957 "uuid": "49a2fc71-16bf-53eb-b3d7-4477702cd667", 00:22:44.957 "is_configured": true, 00:22:44.957 "data_offset": 2048, 00:22:44.957 "data_size": 63488 00:22:44.957 }, 00:22:44.957 { 00:22:44.957 "name": "BaseBdev3", 00:22:44.957 "uuid": "c6ec75d3-d9a0-53b5-bcdb-6b9953600be0", 00:22:44.957 "is_configured": true, 00:22:44.957 "data_offset": 2048, 00:22:44.957 "data_size": 63488 00:22:44.957 } 00:22:44.957 ] 00:22:44.957 }' 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:44.957 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:45.525 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:22:45.525 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:45.525 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:45.525 [2024-10-01 20:21:40.652654] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:22:45.525 [2024-10-01 20:21:40.652700] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:22:45.525 [2024-10-01 20:21:40.656101] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:45.525 [2024-10-01 20:21:40.656166] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:45.525 [2024-10-01 20:21:40.656223] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:22:45.525 [2024-10-01 20:21:40.656239] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:22:45.525 { 00:22:45.525 "results": [ 00:22:45.525 { 00:22:45.525 "job": "raid_bdev1", 00:22:45.525 "core_mask": "0x1", 00:22:45.525 "workload": "randrw", 00:22:45.525 "percentage": 50, 00:22:45.525 "status": "finished", 00:22:45.525 "queue_depth": 1, 00:22:45.525 "io_size": 131072, 00:22:45.525 "runtime": 1.40729, 00:22:45.525 "iops": 9717.257992311464, 00:22:45.525 "mibps": 1214.657249038933, 00:22:45.525 "io_failed": 1, 00:22:45.525 "io_timeout": 0, 00:22:45.525 "avg_latency_us": 144.06049615783456, 00:22:45.525 "min_latency_us": 43.054545454545455, 00:22:45.525 "max_latency_us": 1861.8181818181818 00:22:45.525 } 00:22:45.525 ], 00:22:45.525 "core_count": 1 00:22:45.525 } 00:22:45.525 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:45.525 20:21:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 67885 00:22:45.525 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 67885 ']' 00:22:45.525 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 67885 00:22:45.525 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:22:45.525 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:45.525 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 67885 00:22:45.525 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:22:45.525 killing process with pid 67885 00:22:45.525 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:22:45.525 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 67885' 00:22:45.525 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 67885 00:22:45.525 [2024-10-01 20:21:40.691371] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:22:45.525 20:21:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 67885 00:22:45.784 [2024-10-01 20:21:40.903859] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:22:47.687 20:21:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.Uhwa73Zq1j 00:22:47.687 20:21:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:22:47.687 20:21:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:22:47.687 20:21:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.71 00:22:47.687 20:21:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:22:47.687 20:21:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:22:47.687 20:21:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:22:47.687 20:21:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.71 != \0\.\0\0 ]] 00:22:47.687 00:22:47.687 real 0m5.582s 00:22:47.687 user 0m6.564s 00:22:47.687 sys 0m0.724s 00:22:47.687 ************************************ 00:22:47.687 END TEST raid_write_error_test 00:22:47.687 ************************************ 00:22:47.687 20:21:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:22:47.687 20:21:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:22:47.687 20:21:42 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:22:47.687 20:21:42 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 3 false 00:22:47.687 20:21:42 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:22:47.687 20:21:42 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:22:47.687 20:21:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:22:47.687 ************************************ 00:22:47.687 START TEST raid_state_function_test 00:22:47.687 ************************************ 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 3 false 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=68036 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:22:47.688 Process raid pid: 68036 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 68036' 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 68036 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 68036 ']' 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:47.688 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:47.688 20:21:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:47.947 [2024-10-01 20:21:42.996824] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:22:47.947 [2024-10-01 20:21:42.997055] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:22:47.947 [2024-10-01 20:21:43.175414] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:48.206 [2024-10-01 20:21:43.422150] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:22:48.464 [2024-10-01 20:21:43.633034] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:48.464 [2024-10-01 20:21:43.633098] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:49.031 [2024-10-01 20:21:44.049240] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:22:49.031 [2024-10-01 20:21:44.049309] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:22:49.031 [2024-10-01 20:21:44.049328] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:22:49.031 [2024-10-01 20:21:44.049346] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:22:49.031 [2024-10-01 20:21:44.049356] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:22:49.031 [2024-10-01 20:21:44.049375] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:49.031 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:49.031 "name": "Existed_Raid", 00:22:49.031 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:49.031 "strip_size_kb": 0, 00:22:49.031 "state": "configuring", 00:22:49.031 "raid_level": "raid1", 00:22:49.032 "superblock": false, 00:22:49.032 "num_base_bdevs": 3, 00:22:49.032 "num_base_bdevs_discovered": 0, 00:22:49.032 "num_base_bdevs_operational": 3, 00:22:49.032 "base_bdevs_list": [ 00:22:49.032 { 00:22:49.032 "name": "BaseBdev1", 00:22:49.032 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:49.032 "is_configured": false, 00:22:49.032 "data_offset": 0, 00:22:49.032 "data_size": 0 00:22:49.032 }, 00:22:49.032 { 00:22:49.032 "name": "BaseBdev2", 00:22:49.032 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:49.032 "is_configured": false, 00:22:49.032 "data_offset": 0, 00:22:49.032 "data_size": 0 00:22:49.032 }, 00:22:49.032 { 00:22:49.032 "name": "BaseBdev3", 00:22:49.032 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:49.032 "is_configured": false, 00:22:49.032 "data_offset": 0, 00:22:49.032 "data_size": 0 00:22:49.032 } 00:22:49.032 ] 00:22:49.032 }' 00:22:49.032 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:49.032 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:49.599 [2024-10-01 20:21:44.593302] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:22:49.599 [2024-10-01 20:21:44.593360] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:49.599 [2024-10-01 20:21:44.601276] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:22:49.599 [2024-10-01 20:21:44.601332] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:22:49.599 [2024-10-01 20:21:44.601348] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:22:49.599 [2024-10-01 20:21:44.601365] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:22:49.599 [2024-10-01 20:21:44.601376] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:22:49.599 [2024-10-01 20:21:44.601391] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:49.599 [2024-10-01 20:21:44.647336] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:22:49.599 BaseBdev1 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:49.599 [ 00:22:49.599 { 00:22:49.599 "name": "BaseBdev1", 00:22:49.599 "aliases": [ 00:22:49.599 "1b865335-86eb-4209-bd20-a0a5ad01088e" 00:22:49.599 ], 00:22:49.599 "product_name": "Malloc disk", 00:22:49.599 "block_size": 512, 00:22:49.599 "num_blocks": 65536, 00:22:49.599 "uuid": "1b865335-86eb-4209-bd20-a0a5ad01088e", 00:22:49.599 "assigned_rate_limits": { 00:22:49.599 "rw_ios_per_sec": 0, 00:22:49.599 "rw_mbytes_per_sec": 0, 00:22:49.599 "r_mbytes_per_sec": 0, 00:22:49.599 "w_mbytes_per_sec": 0 00:22:49.599 }, 00:22:49.599 "claimed": true, 00:22:49.599 "claim_type": "exclusive_write", 00:22:49.599 "zoned": false, 00:22:49.599 "supported_io_types": { 00:22:49.599 "read": true, 00:22:49.599 "write": true, 00:22:49.599 "unmap": true, 00:22:49.599 "flush": true, 00:22:49.599 "reset": true, 00:22:49.599 "nvme_admin": false, 00:22:49.599 "nvme_io": false, 00:22:49.599 "nvme_io_md": false, 00:22:49.599 "write_zeroes": true, 00:22:49.599 "zcopy": true, 00:22:49.599 "get_zone_info": false, 00:22:49.599 "zone_management": false, 00:22:49.599 "zone_append": false, 00:22:49.599 "compare": false, 00:22:49.599 "compare_and_write": false, 00:22:49.599 "abort": true, 00:22:49.599 "seek_hole": false, 00:22:49.599 "seek_data": false, 00:22:49.599 "copy": true, 00:22:49.599 "nvme_iov_md": false 00:22:49.599 }, 00:22:49.599 "memory_domains": [ 00:22:49.599 { 00:22:49.599 "dma_device_id": "system", 00:22:49.599 "dma_device_type": 1 00:22:49.599 }, 00:22:49.599 { 00:22:49.599 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:49.599 "dma_device_type": 2 00:22:49.599 } 00:22:49.599 ], 00:22:49.599 "driver_specific": {} 00:22:49.599 } 00:22:49.599 ] 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:49.599 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:49.599 "name": "Existed_Raid", 00:22:49.599 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:49.599 "strip_size_kb": 0, 00:22:49.599 "state": "configuring", 00:22:49.599 "raid_level": "raid1", 00:22:49.599 "superblock": false, 00:22:49.599 "num_base_bdevs": 3, 00:22:49.599 "num_base_bdevs_discovered": 1, 00:22:49.599 "num_base_bdevs_operational": 3, 00:22:49.599 "base_bdevs_list": [ 00:22:49.599 { 00:22:49.599 "name": "BaseBdev1", 00:22:49.599 "uuid": "1b865335-86eb-4209-bd20-a0a5ad01088e", 00:22:49.599 "is_configured": true, 00:22:49.599 "data_offset": 0, 00:22:49.599 "data_size": 65536 00:22:49.599 }, 00:22:49.599 { 00:22:49.599 "name": "BaseBdev2", 00:22:49.599 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:49.599 "is_configured": false, 00:22:49.599 "data_offset": 0, 00:22:49.599 "data_size": 0 00:22:49.599 }, 00:22:49.599 { 00:22:49.600 "name": "BaseBdev3", 00:22:49.600 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:49.600 "is_configured": false, 00:22:49.600 "data_offset": 0, 00:22:49.600 "data_size": 0 00:22:49.600 } 00:22:49.600 ] 00:22:49.600 }' 00:22:49.600 20:21:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:49.600 20:21:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:50.168 [2024-10-01 20:21:45.199550] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:22:50.168 [2024-10-01 20:21:45.199643] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:50.168 [2024-10-01 20:21:45.207603] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:22:50.168 [2024-10-01 20:21:45.210192] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:22:50.168 [2024-10-01 20:21:45.210247] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:22:50.168 [2024-10-01 20:21:45.210266] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:22:50.168 [2024-10-01 20:21:45.210282] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:50.168 "name": "Existed_Raid", 00:22:50.168 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:50.168 "strip_size_kb": 0, 00:22:50.168 "state": "configuring", 00:22:50.168 "raid_level": "raid1", 00:22:50.168 "superblock": false, 00:22:50.168 "num_base_bdevs": 3, 00:22:50.168 "num_base_bdevs_discovered": 1, 00:22:50.168 "num_base_bdevs_operational": 3, 00:22:50.168 "base_bdevs_list": [ 00:22:50.168 { 00:22:50.168 "name": "BaseBdev1", 00:22:50.168 "uuid": "1b865335-86eb-4209-bd20-a0a5ad01088e", 00:22:50.168 "is_configured": true, 00:22:50.168 "data_offset": 0, 00:22:50.168 "data_size": 65536 00:22:50.168 }, 00:22:50.168 { 00:22:50.168 "name": "BaseBdev2", 00:22:50.168 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:50.168 "is_configured": false, 00:22:50.168 "data_offset": 0, 00:22:50.168 "data_size": 0 00:22:50.168 }, 00:22:50.168 { 00:22:50.168 "name": "BaseBdev3", 00:22:50.168 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:50.168 "is_configured": false, 00:22:50.168 "data_offset": 0, 00:22:50.168 "data_size": 0 00:22:50.168 } 00:22:50.168 ] 00:22:50.168 }' 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:50.168 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:50.759 [2024-10-01 20:21:45.734843] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:22:50.759 BaseBdev2 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:50.759 [ 00:22:50.759 { 00:22:50.759 "name": "BaseBdev2", 00:22:50.759 "aliases": [ 00:22:50.759 "7a06d32f-0c1a-464a-8431-cf4934ef67fc" 00:22:50.759 ], 00:22:50.759 "product_name": "Malloc disk", 00:22:50.759 "block_size": 512, 00:22:50.759 "num_blocks": 65536, 00:22:50.759 "uuid": "7a06d32f-0c1a-464a-8431-cf4934ef67fc", 00:22:50.759 "assigned_rate_limits": { 00:22:50.759 "rw_ios_per_sec": 0, 00:22:50.759 "rw_mbytes_per_sec": 0, 00:22:50.759 "r_mbytes_per_sec": 0, 00:22:50.759 "w_mbytes_per_sec": 0 00:22:50.759 }, 00:22:50.759 "claimed": true, 00:22:50.759 "claim_type": "exclusive_write", 00:22:50.759 "zoned": false, 00:22:50.759 "supported_io_types": { 00:22:50.759 "read": true, 00:22:50.759 "write": true, 00:22:50.759 "unmap": true, 00:22:50.759 "flush": true, 00:22:50.759 "reset": true, 00:22:50.759 "nvme_admin": false, 00:22:50.759 "nvme_io": false, 00:22:50.759 "nvme_io_md": false, 00:22:50.759 "write_zeroes": true, 00:22:50.759 "zcopy": true, 00:22:50.759 "get_zone_info": false, 00:22:50.759 "zone_management": false, 00:22:50.759 "zone_append": false, 00:22:50.759 "compare": false, 00:22:50.759 "compare_and_write": false, 00:22:50.759 "abort": true, 00:22:50.759 "seek_hole": false, 00:22:50.759 "seek_data": false, 00:22:50.759 "copy": true, 00:22:50.759 "nvme_iov_md": false 00:22:50.759 }, 00:22:50.759 "memory_domains": [ 00:22:50.759 { 00:22:50.759 "dma_device_id": "system", 00:22:50.759 "dma_device_type": 1 00:22:50.759 }, 00:22:50.759 { 00:22:50.759 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:50.759 "dma_device_type": 2 00:22:50.759 } 00:22:50.759 ], 00:22:50.759 "driver_specific": {} 00:22:50.759 } 00:22:50.759 ] 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:50.759 "name": "Existed_Raid", 00:22:50.759 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:50.759 "strip_size_kb": 0, 00:22:50.759 "state": "configuring", 00:22:50.759 "raid_level": "raid1", 00:22:50.759 "superblock": false, 00:22:50.759 "num_base_bdevs": 3, 00:22:50.759 "num_base_bdevs_discovered": 2, 00:22:50.759 "num_base_bdevs_operational": 3, 00:22:50.759 "base_bdevs_list": [ 00:22:50.759 { 00:22:50.759 "name": "BaseBdev1", 00:22:50.759 "uuid": "1b865335-86eb-4209-bd20-a0a5ad01088e", 00:22:50.759 "is_configured": true, 00:22:50.759 "data_offset": 0, 00:22:50.759 "data_size": 65536 00:22:50.759 }, 00:22:50.759 { 00:22:50.759 "name": "BaseBdev2", 00:22:50.759 "uuid": "7a06d32f-0c1a-464a-8431-cf4934ef67fc", 00:22:50.759 "is_configured": true, 00:22:50.759 "data_offset": 0, 00:22:50.759 "data_size": 65536 00:22:50.759 }, 00:22:50.759 { 00:22:50.759 "name": "BaseBdev3", 00:22:50.759 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:50.759 "is_configured": false, 00:22:50.759 "data_offset": 0, 00:22:50.759 "data_size": 0 00:22:50.759 } 00:22:50.759 ] 00:22:50.759 }' 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:50.759 20:21:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:51.022 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:22:51.022 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:51.022 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:51.282 [2024-10-01 20:21:46.302015] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:22:51.282 [2024-10-01 20:21:46.302091] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:22:51.282 [2024-10-01 20:21:46.302121] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:22:51.282 [2024-10-01 20:21:46.302471] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:22:51.282 [2024-10-01 20:21:46.302701] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:22:51.282 [2024-10-01 20:21:46.302738] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:22:51.282 [2024-10-01 20:21:46.303075] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:51.282 BaseBdev3 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:51.282 [ 00:22:51.282 { 00:22:51.282 "name": "BaseBdev3", 00:22:51.282 "aliases": [ 00:22:51.282 "1ae8d859-871b-4170-9e31-7cd424623de4" 00:22:51.282 ], 00:22:51.282 "product_name": "Malloc disk", 00:22:51.282 "block_size": 512, 00:22:51.282 "num_blocks": 65536, 00:22:51.282 "uuid": "1ae8d859-871b-4170-9e31-7cd424623de4", 00:22:51.282 "assigned_rate_limits": { 00:22:51.282 "rw_ios_per_sec": 0, 00:22:51.282 "rw_mbytes_per_sec": 0, 00:22:51.282 "r_mbytes_per_sec": 0, 00:22:51.282 "w_mbytes_per_sec": 0 00:22:51.282 }, 00:22:51.282 "claimed": true, 00:22:51.282 "claim_type": "exclusive_write", 00:22:51.282 "zoned": false, 00:22:51.282 "supported_io_types": { 00:22:51.282 "read": true, 00:22:51.282 "write": true, 00:22:51.282 "unmap": true, 00:22:51.282 "flush": true, 00:22:51.282 "reset": true, 00:22:51.282 "nvme_admin": false, 00:22:51.282 "nvme_io": false, 00:22:51.282 "nvme_io_md": false, 00:22:51.282 "write_zeroes": true, 00:22:51.282 "zcopy": true, 00:22:51.282 "get_zone_info": false, 00:22:51.282 "zone_management": false, 00:22:51.282 "zone_append": false, 00:22:51.282 "compare": false, 00:22:51.282 "compare_and_write": false, 00:22:51.282 "abort": true, 00:22:51.282 "seek_hole": false, 00:22:51.282 "seek_data": false, 00:22:51.282 "copy": true, 00:22:51.282 "nvme_iov_md": false 00:22:51.282 }, 00:22:51.282 "memory_domains": [ 00:22:51.282 { 00:22:51.282 "dma_device_id": "system", 00:22:51.282 "dma_device_type": 1 00:22:51.282 }, 00:22:51.282 { 00:22:51.282 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:51.282 "dma_device_type": 2 00:22:51.282 } 00:22:51.282 ], 00:22:51.282 "driver_specific": {} 00:22:51.282 } 00:22:51.282 ] 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:51.282 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:51.283 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:51.283 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:51.283 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:51.283 "name": "Existed_Raid", 00:22:51.283 "uuid": "186dc5d4-4714-4fb5-ad45-879e0321f87e", 00:22:51.283 "strip_size_kb": 0, 00:22:51.283 "state": "online", 00:22:51.283 "raid_level": "raid1", 00:22:51.283 "superblock": false, 00:22:51.283 "num_base_bdevs": 3, 00:22:51.283 "num_base_bdevs_discovered": 3, 00:22:51.283 "num_base_bdevs_operational": 3, 00:22:51.283 "base_bdevs_list": [ 00:22:51.283 { 00:22:51.283 "name": "BaseBdev1", 00:22:51.283 "uuid": "1b865335-86eb-4209-bd20-a0a5ad01088e", 00:22:51.283 "is_configured": true, 00:22:51.283 "data_offset": 0, 00:22:51.283 "data_size": 65536 00:22:51.283 }, 00:22:51.283 { 00:22:51.283 "name": "BaseBdev2", 00:22:51.283 "uuid": "7a06d32f-0c1a-464a-8431-cf4934ef67fc", 00:22:51.283 "is_configured": true, 00:22:51.283 "data_offset": 0, 00:22:51.283 "data_size": 65536 00:22:51.283 }, 00:22:51.283 { 00:22:51.283 "name": "BaseBdev3", 00:22:51.283 "uuid": "1ae8d859-871b-4170-9e31-7cd424623de4", 00:22:51.283 "is_configured": true, 00:22:51.283 "data_offset": 0, 00:22:51.283 "data_size": 65536 00:22:51.283 } 00:22:51.283 ] 00:22:51.283 }' 00:22:51.283 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:51.283 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:51.851 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:22:51.851 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:22:51.851 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:22:51.851 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:22:51.851 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:22:51.851 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:22:51.851 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:22:51.851 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:22:51.851 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:51.851 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:51.851 [2024-10-01 20:21:46.846648] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:51.851 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:51.851 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:22:51.851 "name": "Existed_Raid", 00:22:51.851 "aliases": [ 00:22:51.851 "186dc5d4-4714-4fb5-ad45-879e0321f87e" 00:22:51.851 ], 00:22:51.851 "product_name": "Raid Volume", 00:22:51.851 "block_size": 512, 00:22:51.851 "num_blocks": 65536, 00:22:51.851 "uuid": "186dc5d4-4714-4fb5-ad45-879e0321f87e", 00:22:51.851 "assigned_rate_limits": { 00:22:51.851 "rw_ios_per_sec": 0, 00:22:51.851 "rw_mbytes_per_sec": 0, 00:22:51.851 "r_mbytes_per_sec": 0, 00:22:51.851 "w_mbytes_per_sec": 0 00:22:51.851 }, 00:22:51.851 "claimed": false, 00:22:51.851 "zoned": false, 00:22:51.851 "supported_io_types": { 00:22:51.851 "read": true, 00:22:51.851 "write": true, 00:22:51.851 "unmap": false, 00:22:51.851 "flush": false, 00:22:51.851 "reset": true, 00:22:51.851 "nvme_admin": false, 00:22:51.851 "nvme_io": false, 00:22:51.851 "nvme_io_md": false, 00:22:51.851 "write_zeroes": true, 00:22:51.851 "zcopy": false, 00:22:51.851 "get_zone_info": false, 00:22:51.851 "zone_management": false, 00:22:51.851 "zone_append": false, 00:22:51.851 "compare": false, 00:22:51.851 "compare_and_write": false, 00:22:51.851 "abort": false, 00:22:51.851 "seek_hole": false, 00:22:51.851 "seek_data": false, 00:22:51.851 "copy": false, 00:22:51.851 "nvme_iov_md": false 00:22:51.851 }, 00:22:51.851 "memory_domains": [ 00:22:51.851 { 00:22:51.851 "dma_device_id": "system", 00:22:51.851 "dma_device_type": 1 00:22:51.851 }, 00:22:51.851 { 00:22:51.851 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:51.851 "dma_device_type": 2 00:22:51.851 }, 00:22:51.851 { 00:22:51.851 "dma_device_id": "system", 00:22:51.851 "dma_device_type": 1 00:22:51.851 }, 00:22:51.851 { 00:22:51.851 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:51.851 "dma_device_type": 2 00:22:51.851 }, 00:22:51.851 { 00:22:51.851 "dma_device_id": "system", 00:22:51.851 "dma_device_type": 1 00:22:51.851 }, 00:22:51.851 { 00:22:51.852 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:51.852 "dma_device_type": 2 00:22:51.852 } 00:22:51.852 ], 00:22:51.852 "driver_specific": { 00:22:51.852 "raid": { 00:22:51.852 "uuid": "186dc5d4-4714-4fb5-ad45-879e0321f87e", 00:22:51.852 "strip_size_kb": 0, 00:22:51.852 "state": "online", 00:22:51.852 "raid_level": "raid1", 00:22:51.852 "superblock": false, 00:22:51.852 "num_base_bdevs": 3, 00:22:51.852 "num_base_bdevs_discovered": 3, 00:22:51.852 "num_base_bdevs_operational": 3, 00:22:51.852 "base_bdevs_list": [ 00:22:51.852 { 00:22:51.852 "name": "BaseBdev1", 00:22:51.852 "uuid": "1b865335-86eb-4209-bd20-a0a5ad01088e", 00:22:51.852 "is_configured": true, 00:22:51.852 "data_offset": 0, 00:22:51.852 "data_size": 65536 00:22:51.852 }, 00:22:51.852 { 00:22:51.852 "name": "BaseBdev2", 00:22:51.852 "uuid": "7a06d32f-0c1a-464a-8431-cf4934ef67fc", 00:22:51.852 "is_configured": true, 00:22:51.852 "data_offset": 0, 00:22:51.852 "data_size": 65536 00:22:51.852 }, 00:22:51.852 { 00:22:51.852 "name": "BaseBdev3", 00:22:51.852 "uuid": "1ae8d859-871b-4170-9e31-7cd424623de4", 00:22:51.852 "is_configured": true, 00:22:51.852 "data_offset": 0, 00:22:51.852 "data_size": 65536 00:22:51.852 } 00:22:51.852 ] 00:22:51.852 } 00:22:51.852 } 00:22:51.852 }' 00:22:51.852 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:22:51.852 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:22:51.852 BaseBdev2 00:22:51.852 BaseBdev3' 00:22:51.852 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:51.852 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:22:51.852 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:51.852 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:22:51.852 20:21:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:51.852 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:51.852 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:51.852 20:21:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:51.852 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:51.852 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:51.852 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:51.852 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:22:51.852 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:51.852 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:51.852 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:51.852 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:51.852 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:51.852 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:51.852 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:51.852 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:22:51.852 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:51.852 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:51.852 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:52.111 [2024-10-01 20:21:47.146339] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:52.111 "name": "Existed_Raid", 00:22:52.111 "uuid": "186dc5d4-4714-4fb5-ad45-879e0321f87e", 00:22:52.111 "strip_size_kb": 0, 00:22:52.111 "state": "online", 00:22:52.111 "raid_level": "raid1", 00:22:52.111 "superblock": false, 00:22:52.111 "num_base_bdevs": 3, 00:22:52.111 "num_base_bdevs_discovered": 2, 00:22:52.111 "num_base_bdevs_operational": 2, 00:22:52.111 "base_bdevs_list": [ 00:22:52.111 { 00:22:52.111 "name": null, 00:22:52.111 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:52.111 "is_configured": false, 00:22:52.111 "data_offset": 0, 00:22:52.111 "data_size": 65536 00:22:52.111 }, 00:22:52.111 { 00:22:52.111 "name": "BaseBdev2", 00:22:52.111 "uuid": "7a06d32f-0c1a-464a-8431-cf4934ef67fc", 00:22:52.111 "is_configured": true, 00:22:52.111 "data_offset": 0, 00:22:52.111 "data_size": 65536 00:22:52.111 }, 00:22:52.111 { 00:22:52.111 "name": "BaseBdev3", 00:22:52.111 "uuid": "1ae8d859-871b-4170-9e31-7cd424623de4", 00:22:52.111 "is_configured": true, 00:22:52.111 "data_offset": 0, 00:22:52.111 "data_size": 65536 00:22:52.111 } 00:22:52.111 ] 00:22:52.111 }' 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:52.111 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:52.679 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:22:52.679 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:22:52.679 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:52.679 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:22:52.679 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:52.679 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:52.679 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:52.679 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:22:52.680 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:22:52.680 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:22:52.680 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:52.680 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:52.680 [2024-10-01 20:21:47.796887] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:22:52.680 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:52.680 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:22:52.680 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:22:52.680 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:52.680 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:22:52.680 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:52.680 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:52.680 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:52.939 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:22:52.939 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:22:52.939 20:21:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:22:52.939 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:52.939 20:21:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:52.939 [2024-10-01 20:21:47.952012] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:22:52.939 [2024-10-01 20:21:47.952311] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:22:52.939 [2024-10-01 20:21:48.040239] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:52.939 [2024-10-01 20:21:48.040324] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:22:52.939 [2024-10-01 20:21:48.040347] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:52.939 BaseBdev2 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:52.939 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:52.939 [ 00:22:52.939 { 00:22:52.939 "name": "BaseBdev2", 00:22:52.939 "aliases": [ 00:22:52.939 "2491c869-b47f-4966-9f63-31131e05b1c3" 00:22:52.939 ], 00:22:52.939 "product_name": "Malloc disk", 00:22:52.939 "block_size": 512, 00:22:52.939 "num_blocks": 65536, 00:22:52.939 "uuid": "2491c869-b47f-4966-9f63-31131e05b1c3", 00:22:52.939 "assigned_rate_limits": { 00:22:52.939 "rw_ios_per_sec": 0, 00:22:52.939 "rw_mbytes_per_sec": 0, 00:22:52.939 "r_mbytes_per_sec": 0, 00:22:52.939 "w_mbytes_per_sec": 0 00:22:52.939 }, 00:22:52.939 "claimed": false, 00:22:52.939 "zoned": false, 00:22:52.939 "supported_io_types": { 00:22:52.939 "read": true, 00:22:52.939 "write": true, 00:22:52.939 "unmap": true, 00:22:52.939 "flush": true, 00:22:52.939 "reset": true, 00:22:52.939 "nvme_admin": false, 00:22:52.939 "nvme_io": false, 00:22:52.939 "nvme_io_md": false, 00:22:52.939 "write_zeroes": true, 00:22:52.939 "zcopy": true, 00:22:52.939 "get_zone_info": false, 00:22:52.939 "zone_management": false, 00:22:52.939 "zone_append": false, 00:22:52.939 "compare": false, 00:22:52.939 "compare_and_write": false, 00:22:52.939 "abort": true, 00:22:52.939 "seek_hole": false, 00:22:52.939 "seek_data": false, 00:22:52.939 "copy": true, 00:22:52.939 "nvme_iov_md": false 00:22:52.939 }, 00:22:52.939 "memory_domains": [ 00:22:52.939 { 00:22:52.939 "dma_device_id": "system", 00:22:52.939 "dma_device_type": 1 00:22:52.939 }, 00:22:52.939 { 00:22:52.940 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:52.940 "dma_device_type": 2 00:22:52.940 } 00:22:52.940 ], 00:22:52.940 "driver_specific": {} 00:22:52.940 } 00:22:52.940 ] 00:22:52.940 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:52.940 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:22:52.940 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:22:52.940 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:22:52.940 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:22:52.940 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:52.940 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:53.199 BaseBdev3 00:22:53.199 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:53.199 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:22:53.199 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:22:53.199 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:53.199 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:22:53.199 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:53.199 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:53.199 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:53.199 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:53.199 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:53.199 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:53.199 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:22:53.199 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:53.199 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:53.199 [ 00:22:53.199 { 00:22:53.199 "name": "BaseBdev3", 00:22:53.199 "aliases": [ 00:22:53.199 "e90ddc37-b7ef-4556-8097-36986a3a88e3" 00:22:53.199 ], 00:22:53.199 "product_name": "Malloc disk", 00:22:53.199 "block_size": 512, 00:22:53.199 "num_blocks": 65536, 00:22:53.199 "uuid": "e90ddc37-b7ef-4556-8097-36986a3a88e3", 00:22:53.199 "assigned_rate_limits": { 00:22:53.200 "rw_ios_per_sec": 0, 00:22:53.200 "rw_mbytes_per_sec": 0, 00:22:53.200 "r_mbytes_per_sec": 0, 00:22:53.200 "w_mbytes_per_sec": 0 00:22:53.200 }, 00:22:53.200 "claimed": false, 00:22:53.200 "zoned": false, 00:22:53.200 "supported_io_types": { 00:22:53.200 "read": true, 00:22:53.200 "write": true, 00:22:53.200 "unmap": true, 00:22:53.200 "flush": true, 00:22:53.200 "reset": true, 00:22:53.200 "nvme_admin": false, 00:22:53.200 "nvme_io": false, 00:22:53.200 "nvme_io_md": false, 00:22:53.200 "write_zeroes": true, 00:22:53.200 "zcopy": true, 00:22:53.200 "get_zone_info": false, 00:22:53.200 "zone_management": false, 00:22:53.200 "zone_append": false, 00:22:53.200 "compare": false, 00:22:53.200 "compare_and_write": false, 00:22:53.200 "abort": true, 00:22:53.200 "seek_hole": false, 00:22:53.200 "seek_data": false, 00:22:53.200 "copy": true, 00:22:53.200 "nvme_iov_md": false 00:22:53.200 }, 00:22:53.200 "memory_domains": [ 00:22:53.200 { 00:22:53.200 "dma_device_id": "system", 00:22:53.200 "dma_device_type": 1 00:22:53.200 }, 00:22:53.200 { 00:22:53.200 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:53.200 "dma_device_type": 2 00:22:53.200 } 00:22:53.200 ], 00:22:53.200 "driver_specific": {} 00:22:53.200 } 00:22:53.200 ] 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:53.200 [2024-10-01 20:21:48.257908] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:22:53.200 [2024-10-01 20:21:48.257978] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:22:53.200 [2024-10-01 20:21:48.258007] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:22:53.200 [2024-10-01 20:21:48.260613] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:53.200 "name": "Existed_Raid", 00:22:53.200 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:53.200 "strip_size_kb": 0, 00:22:53.200 "state": "configuring", 00:22:53.200 "raid_level": "raid1", 00:22:53.200 "superblock": false, 00:22:53.200 "num_base_bdevs": 3, 00:22:53.200 "num_base_bdevs_discovered": 2, 00:22:53.200 "num_base_bdevs_operational": 3, 00:22:53.200 "base_bdevs_list": [ 00:22:53.200 { 00:22:53.200 "name": "BaseBdev1", 00:22:53.200 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:53.200 "is_configured": false, 00:22:53.200 "data_offset": 0, 00:22:53.200 "data_size": 0 00:22:53.200 }, 00:22:53.200 { 00:22:53.200 "name": "BaseBdev2", 00:22:53.200 "uuid": "2491c869-b47f-4966-9f63-31131e05b1c3", 00:22:53.200 "is_configured": true, 00:22:53.200 "data_offset": 0, 00:22:53.200 "data_size": 65536 00:22:53.200 }, 00:22:53.200 { 00:22:53.200 "name": "BaseBdev3", 00:22:53.200 "uuid": "e90ddc37-b7ef-4556-8097-36986a3a88e3", 00:22:53.200 "is_configured": true, 00:22:53.200 "data_offset": 0, 00:22:53.200 "data_size": 65536 00:22:53.200 } 00:22:53.200 ] 00:22:53.200 }' 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:53.200 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:53.767 [2024-10-01 20:21:48.770073] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:53.767 "name": "Existed_Raid", 00:22:53.767 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:53.767 "strip_size_kb": 0, 00:22:53.767 "state": "configuring", 00:22:53.767 "raid_level": "raid1", 00:22:53.767 "superblock": false, 00:22:53.767 "num_base_bdevs": 3, 00:22:53.767 "num_base_bdevs_discovered": 1, 00:22:53.767 "num_base_bdevs_operational": 3, 00:22:53.767 "base_bdevs_list": [ 00:22:53.767 { 00:22:53.767 "name": "BaseBdev1", 00:22:53.767 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:53.767 "is_configured": false, 00:22:53.767 "data_offset": 0, 00:22:53.767 "data_size": 0 00:22:53.767 }, 00:22:53.767 { 00:22:53.767 "name": null, 00:22:53.767 "uuid": "2491c869-b47f-4966-9f63-31131e05b1c3", 00:22:53.767 "is_configured": false, 00:22:53.767 "data_offset": 0, 00:22:53.767 "data_size": 65536 00:22:53.767 }, 00:22:53.767 { 00:22:53.767 "name": "BaseBdev3", 00:22:53.767 "uuid": "e90ddc37-b7ef-4556-8097-36986a3a88e3", 00:22:53.767 "is_configured": true, 00:22:53.767 "data_offset": 0, 00:22:53.767 "data_size": 65536 00:22:53.767 } 00:22:53.767 ] 00:22:53.767 }' 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:53.767 20:21:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:54.138 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:54.138 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:22:54.138 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:54.138 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:54.138 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:54.138 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:22:54.138 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:22:54.138 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:54.138 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:54.138 [2024-10-01 20:21:49.372671] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:22:54.138 BaseBdev1 00:22:54.139 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:54.139 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:22:54.139 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:22:54.139 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:54.139 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:22:54.139 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:54.139 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:54.139 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:54.139 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:54.139 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:54.139 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:54.139 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:22:54.139 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:54.139 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:54.398 [ 00:22:54.398 { 00:22:54.398 "name": "BaseBdev1", 00:22:54.398 "aliases": [ 00:22:54.398 "c4b109f3-8702-4b41-88d5-1ea0b7448817" 00:22:54.398 ], 00:22:54.398 "product_name": "Malloc disk", 00:22:54.398 "block_size": 512, 00:22:54.398 "num_blocks": 65536, 00:22:54.398 "uuid": "c4b109f3-8702-4b41-88d5-1ea0b7448817", 00:22:54.398 "assigned_rate_limits": { 00:22:54.398 "rw_ios_per_sec": 0, 00:22:54.398 "rw_mbytes_per_sec": 0, 00:22:54.398 "r_mbytes_per_sec": 0, 00:22:54.398 "w_mbytes_per_sec": 0 00:22:54.398 }, 00:22:54.398 "claimed": true, 00:22:54.398 "claim_type": "exclusive_write", 00:22:54.398 "zoned": false, 00:22:54.398 "supported_io_types": { 00:22:54.398 "read": true, 00:22:54.398 "write": true, 00:22:54.398 "unmap": true, 00:22:54.398 "flush": true, 00:22:54.398 "reset": true, 00:22:54.398 "nvme_admin": false, 00:22:54.398 "nvme_io": false, 00:22:54.398 "nvme_io_md": false, 00:22:54.398 "write_zeroes": true, 00:22:54.398 "zcopy": true, 00:22:54.398 "get_zone_info": false, 00:22:54.398 "zone_management": false, 00:22:54.398 "zone_append": false, 00:22:54.398 "compare": false, 00:22:54.398 "compare_and_write": false, 00:22:54.398 "abort": true, 00:22:54.398 "seek_hole": false, 00:22:54.398 "seek_data": false, 00:22:54.398 "copy": true, 00:22:54.398 "nvme_iov_md": false 00:22:54.398 }, 00:22:54.398 "memory_domains": [ 00:22:54.398 { 00:22:54.398 "dma_device_id": "system", 00:22:54.398 "dma_device_type": 1 00:22:54.398 }, 00:22:54.398 { 00:22:54.398 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:54.398 "dma_device_type": 2 00:22:54.398 } 00:22:54.398 ], 00:22:54.398 "driver_specific": {} 00:22:54.398 } 00:22:54.398 ] 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:54.398 "name": "Existed_Raid", 00:22:54.398 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:54.398 "strip_size_kb": 0, 00:22:54.398 "state": "configuring", 00:22:54.398 "raid_level": "raid1", 00:22:54.398 "superblock": false, 00:22:54.398 "num_base_bdevs": 3, 00:22:54.398 "num_base_bdevs_discovered": 2, 00:22:54.398 "num_base_bdevs_operational": 3, 00:22:54.398 "base_bdevs_list": [ 00:22:54.398 { 00:22:54.398 "name": "BaseBdev1", 00:22:54.398 "uuid": "c4b109f3-8702-4b41-88d5-1ea0b7448817", 00:22:54.398 "is_configured": true, 00:22:54.398 "data_offset": 0, 00:22:54.398 "data_size": 65536 00:22:54.398 }, 00:22:54.398 { 00:22:54.398 "name": null, 00:22:54.398 "uuid": "2491c869-b47f-4966-9f63-31131e05b1c3", 00:22:54.398 "is_configured": false, 00:22:54.398 "data_offset": 0, 00:22:54.398 "data_size": 65536 00:22:54.398 }, 00:22:54.398 { 00:22:54.398 "name": "BaseBdev3", 00:22:54.398 "uuid": "e90ddc37-b7ef-4556-8097-36986a3a88e3", 00:22:54.398 "is_configured": true, 00:22:54.398 "data_offset": 0, 00:22:54.398 "data_size": 65536 00:22:54.398 } 00:22:54.398 ] 00:22:54.398 }' 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:54.398 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:54.964 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:54.964 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:22:54.964 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:54.964 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:54.964 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:54.964 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:22:54.964 20:21:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:22:54.964 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:54.964 20:21:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:54.964 [2024-10-01 20:21:50.000950] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:22:54.964 20:21:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:54.964 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:22:54.964 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:54.964 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:54.964 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:22:54.964 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:22:54.964 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:54.964 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:54.964 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:54.964 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:54.964 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:54.964 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:54.964 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:54.964 20:21:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:54.964 20:21:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:54.965 20:21:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:54.965 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:54.965 "name": "Existed_Raid", 00:22:54.965 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:54.965 "strip_size_kb": 0, 00:22:54.965 "state": "configuring", 00:22:54.965 "raid_level": "raid1", 00:22:54.965 "superblock": false, 00:22:54.965 "num_base_bdevs": 3, 00:22:54.965 "num_base_bdevs_discovered": 1, 00:22:54.965 "num_base_bdevs_operational": 3, 00:22:54.965 "base_bdevs_list": [ 00:22:54.965 { 00:22:54.965 "name": "BaseBdev1", 00:22:54.965 "uuid": "c4b109f3-8702-4b41-88d5-1ea0b7448817", 00:22:54.965 "is_configured": true, 00:22:54.965 "data_offset": 0, 00:22:54.965 "data_size": 65536 00:22:54.965 }, 00:22:54.965 { 00:22:54.965 "name": null, 00:22:54.965 "uuid": "2491c869-b47f-4966-9f63-31131e05b1c3", 00:22:54.965 "is_configured": false, 00:22:54.965 "data_offset": 0, 00:22:54.965 "data_size": 65536 00:22:54.965 }, 00:22:54.965 { 00:22:54.965 "name": null, 00:22:54.965 "uuid": "e90ddc37-b7ef-4556-8097-36986a3a88e3", 00:22:54.965 "is_configured": false, 00:22:54.965 "data_offset": 0, 00:22:54.965 "data_size": 65536 00:22:54.965 } 00:22:54.965 ] 00:22:54.965 }' 00:22:54.965 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:54.965 20:21:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:55.535 [2024-10-01 20:21:50.629132] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:55.535 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:55.535 "name": "Existed_Raid", 00:22:55.535 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:55.535 "strip_size_kb": 0, 00:22:55.535 "state": "configuring", 00:22:55.535 "raid_level": "raid1", 00:22:55.535 "superblock": false, 00:22:55.535 "num_base_bdevs": 3, 00:22:55.536 "num_base_bdevs_discovered": 2, 00:22:55.536 "num_base_bdevs_operational": 3, 00:22:55.536 "base_bdevs_list": [ 00:22:55.536 { 00:22:55.536 "name": "BaseBdev1", 00:22:55.536 "uuid": "c4b109f3-8702-4b41-88d5-1ea0b7448817", 00:22:55.536 "is_configured": true, 00:22:55.536 "data_offset": 0, 00:22:55.536 "data_size": 65536 00:22:55.536 }, 00:22:55.536 { 00:22:55.536 "name": null, 00:22:55.536 "uuid": "2491c869-b47f-4966-9f63-31131e05b1c3", 00:22:55.536 "is_configured": false, 00:22:55.536 "data_offset": 0, 00:22:55.536 "data_size": 65536 00:22:55.536 }, 00:22:55.536 { 00:22:55.536 "name": "BaseBdev3", 00:22:55.536 "uuid": "e90ddc37-b7ef-4556-8097-36986a3a88e3", 00:22:55.536 "is_configured": true, 00:22:55.536 "data_offset": 0, 00:22:55.536 "data_size": 65536 00:22:55.536 } 00:22:55.536 ] 00:22:55.536 }' 00:22:55.536 20:21:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:55.536 20:21:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:56.103 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:22:56.103 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:56.103 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:56.103 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:56.103 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:56.103 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:22:56.103 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:22:56.103 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:56.103 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:56.104 [2024-10-01 20:21:51.169390] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:56.104 "name": "Existed_Raid", 00:22:56.104 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:56.104 "strip_size_kb": 0, 00:22:56.104 "state": "configuring", 00:22:56.104 "raid_level": "raid1", 00:22:56.104 "superblock": false, 00:22:56.104 "num_base_bdevs": 3, 00:22:56.104 "num_base_bdevs_discovered": 1, 00:22:56.104 "num_base_bdevs_operational": 3, 00:22:56.104 "base_bdevs_list": [ 00:22:56.104 { 00:22:56.104 "name": null, 00:22:56.104 "uuid": "c4b109f3-8702-4b41-88d5-1ea0b7448817", 00:22:56.104 "is_configured": false, 00:22:56.104 "data_offset": 0, 00:22:56.104 "data_size": 65536 00:22:56.104 }, 00:22:56.104 { 00:22:56.104 "name": null, 00:22:56.104 "uuid": "2491c869-b47f-4966-9f63-31131e05b1c3", 00:22:56.104 "is_configured": false, 00:22:56.104 "data_offset": 0, 00:22:56.104 "data_size": 65536 00:22:56.104 }, 00:22:56.104 { 00:22:56.104 "name": "BaseBdev3", 00:22:56.104 "uuid": "e90ddc37-b7ef-4556-8097-36986a3a88e3", 00:22:56.104 "is_configured": true, 00:22:56.104 "data_offset": 0, 00:22:56.104 "data_size": 65536 00:22:56.104 } 00:22:56.104 ] 00:22:56.104 }' 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:56.104 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:56.670 [2024-10-01 20:21:51.837806] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:56.670 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:56.671 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:56.671 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:56.671 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:56.671 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:56.671 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:56.671 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:56.671 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:56.671 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:56.671 "name": "Existed_Raid", 00:22:56.671 "uuid": "00000000-0000-0000-0000-000000000000", 00:22:56.671 "strip_size_kb": 0, 00:22:56.671 "state": "configuring", 00:22:56.671 "raid_level": "raid1", 00:22:56.671 "superblock": false, 00:22:56.671 "num_base_bdevs": 3, 00:22:56.671 "num_base_bdevs_discovered": 2, 00:22:56.671 "num_base_bdevs_operational": 3, 00:22:56.671 "base_bdevs_list": [ 00:22:56.671 { 00:22:56.671 "name": null, 00:22:56.671 "uuid": "c4b109f3-8702-4b41-88d5-1ea0b7448817", 00:22:56.671 "is_configured": false, 00:22:56.671 "data_offset": 0, 00:22:56.671 "data_size": 65536 00:22:56.671 }, 00:22:56.671 { 00:22:56.671 "name": "BaseBdev2", 00:22:56.671 "uuid": "2491c869-b47f-4966-9f63-31131e05b1c3", 00:22:56.671 "is_configured": true, 00:22:56.671 "data_offset": 0, 00:22:56.671 "data_size": 65536 00:22:56.671 }, 00:22:56.671 { 00:22:56.671 "name": "BaseBdev3", 00:22:56.671 "uuid": "e90ddc37-b7ef-4556-8097-36986a3a88e3", 00:22:56.671 "is_configured": true, 00:22:56.671 "data_offset": 0, 00:22:56.671 "data_size": 65536 00:22:56.671 } 00:22:56.671 ] 00:22:56.671 }' 00:22:56.671 20:21:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:56.671 20:21:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:57.237 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:57.238 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:22:57.238 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:57.238 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:57.238 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:57.238 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:22:57.238 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:57.238 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:22:57.238 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:57.238 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:57.238 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:57.496 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u c4b109f3-8702-4b41-88d5-1ea0b7448817 00:22:57.496 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:57.496 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:57.496 [2024-10-01 20:21:52.545658] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:22:57.496 NewBaseBdev 00:22:57.496 [2024-10-01 20:21:52.546033] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:22:57.496 [2024-10-01 20:21:52.546058] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:22:57.496 [2024-10-01 20:21:52.546401] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:22:57.496 [2024-10-01 20:21:52.546637] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:22:57.496 [2024-10-01 20:21:52.546660] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:22:57.496 [2024-10-01 20:21:52.547024] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:22:57.496 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:57.496 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:22:57.496 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:22:57.496 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:22:57.496 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:22:57.496 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:22:57.496 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:22:57.496 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:22:57.496 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:57.496 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:57.496 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:57.496 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:57.497 [ 00:22:57.497 { 00:22:57.497 "name": "NewBaseBdev", 00:22:57.497 "aliases": [ 00:22:57.497 "c4b109f3-8702-4b41-88d5-1ea0b7448817" 00:22:57.497 ], 00:22:57.497 "product_name": "Malloc disk", 00:22:57.497 "block_size": 512, 00:22:57.497 "num_blocks": 65536, 00:22:57.497 "uuid": "c4b109f3-8702-4b41-88d5-1ea0b7448817", 00:22:57.497 "assigned_rate_limits": { 00:22:57.497 "rw_ios_per_sec": 0, 00:22:57.497 "rw_mbytes_per_sec": 0, 00:22:57.497 "r_mbytes_per_sec": 0, 00:22:57.497 "w_mbytes_per_sec": 0 00:22:57.497 }, 00:22:57.497 "claimed": true, 00:22:57.497 "claim_type": "exclusive_write", 00:22:57.497 "zoned": false, 00:22:57.497 "supported_io_types": { 00:22:57.497 "read": true, 00:22:57.497 "write": true, 00:22:57.497 "unmap": true, 00:22:57.497 "flush": true, 00:22:57.497 "reset": true, 00:22:57.497 "nvme_admin": false, 00:22:57.497 "nvme_io": false, 00:22:57.497 "nvme_io_md": false, 00:22:57.497 "write_zeroes": true, 00:22:57.497 "zcopy": true, 00:22:57.497 "get_zone_info": false, 00:22:57.497 "zone_management": false, 00:22:57.497 "zone_append": false, 00:22:57.497 "compare": false, 00:22:57.497 "compare_and_write": false, 00:22:57.497 "abort": true, 00:22:57.497 "seek_hole": false, 00:22:57.497 "seek_data": false, 00:22:57.497 "copy": true, 00:22:57.497 "nvme_iov_md": false 00:22:57.497 }, 00:22:57.497 "memory_domains": [ 00:22:57.497 { 00:22:57.497 "dma_device_id": "system", 00:22:57.497 "dma_device_type": 1 00:22:57.497 }, 00:22:57.497 { 00:22:57.497 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:57.497 "dma_device_type": 2 00:22:57.497 } 00:22:57.497 ], 00:22:57.497 "driver_specific": {} 00:22:57.497 } 00:22:57.497 ] 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:22:57.497 "name": "Existed_Raid", 00:22:57.497 "uuid": "79bf6545-3395-4966-9ef3-90dc42b3e780", 00:22:57.497 "strip_size_kb": 0, 00:22:57.497 "state": "online", 00:22:57.497 "raid_level": "raid1", 00:22:57.497 "superblock": false, 00:22:57.497 "num_base_bdevs": 3, 00:22:57.497 "num_base_bdevs_discovered": 3, 00:22:57.497 "num_base_bdevs_operational": 3, 00:22:57.497 "base_bdevs_list": [ 00:22:57.497 { 00:22:57.497 "name": "NewBaseBdev", 00:22:57.497 "uuid": "c4b109f3-8702-4b41-88d5-1ea0b7448817", 00:22:57.497 "is_configured": true, 00:22:57.497 "data_offset": 0, 00:22:57.497 "data_size": 65536 00:22:57.497 }, 00:22:57.497 { 00:22:57.497 "name": "BaseBdev2", 00:22:57.497 "uuid": "2491c869-b47f-4966-9f63-31131e05b1c3", 00:22:57.497 "is_configured": true, 00:22:57.497 "data_offset": 0, 00:22:57.497 "data_size": 65536 00:22:57.497 }, 00:22:57.497 { 00:22:57.497 "name": "BaseBdev3", 00:22:57.497 "uuid": "e90ddc37-b7ef-4556-8097-36986a3a88e3", 00:22:57.497 "is_configured": true, 00:22:57.497 "data_offset": 0, 00:22:57.497 "data_size": 65536 00:22:57.497 } 00:22:57.497 ] 00:22:57.497 }' 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:22:57.497 20:21:52 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:58.064 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:22:58.064 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:22:58.064 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:22:58.064 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:22:58.064 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:22:58.064 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:22:58.064 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:22:58.064 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:22:58.064 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:58.064 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:58.064 [2024-10-01 20:21:53.186410] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:22:58.064 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:58.064 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:22:58.064 "name": "Existed_Raid", 00:22:58.064 "aliases": [ 00:22:58.064 "79bf6545-3395-4966-9ef3-90dc42b3e780" 00:22:58.064 ], 00:22:58.064 "product_name": "Raid Volume", 00:22:58.064 "block_size": 512, 00:22:58.064 "num_blocks": 65536, 00:22:58.064 "uuid": "79bf6545-3395-4966-9ef3-90dc42b3e780", 00:22:58.064 "assigned_rate_limits": { 00:22:58.064 "rw_ios_per_sec": 0, 00:22:58.064 "rw_mbytes_per_sec": 0, 00:22:58.064 "r_mbytes_per_sec": 0, 00:22:58.064 "w_mbytes_per_sec": 0 00:22:58.064 }, 00:22:58.064 "claimed": false, 00:22:58.064 "zoned": false, 00:22:58.064 "supported_io_types": { 00:22:58.064 "read": true, 00:22:58.064 "write": true, 00:22:58.064 "unmap": false, 00:22:58.064 "flush": false, 00:22:58.064 "reset": true, 00:22:58.064 "nvme_admin": false, 00:22:58.064 "nvme_io": false, 00:22:58.064 "nvme_io_md": false, 00:22:58.064 "write_zeroes": true, 00:22:58.064 "zcopy": false, 00:22:58.064 "get_zone_info": false, 00:22:58.064 "zone_management": false, 00:22:58.064 "zone_append": false, 00:22:58.064 "compare": false, 00:22:58.064 "compare_and_write": false, 00:22:58.064 "abort": false, 00:22:58.064 "seek_hole": false, 00:22:58.064 "seek_data": false, 00:22:58.064 "copy": false, 00:22:58.064 "nvme_iov_md": false 00:22:58.064 }, 00:22:58.064 "memory_domains": [ 00:22:58.064 { 00:22:58.064 "dma_device_id": "system", 00:22:58.064 "dma_device_type": 1 00:22:58.064 }, 00:22:58.064 { 00:22:58.064 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:58.064 "dma_device_type": 2 00:22:58.064 }, 00:22:58.064 { 00:22:58.064 "dma_device_id": "system", 00:22:58.064 "dma_device_type": 1 00:22:58.064 }, 00:22:58.064 { 00:22:58.064 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:58.064 "dma_device_type": 2 00:22:58.064 }, 00:22:58.064 { 00:22:58.064 "dma_device_id": "system", 00:22:58.064 "dma_device_type": 1 00:22:58.064 }, 00:22:58.064 { 00:22:58.065 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:22:58.065 "dma_device_type": 2 00:22:58.065 } 00:22:58.065 ], 00:22:58.065 "driver_specific": { 00:22:58.065 "raid": { 00:22:58.065 "uuid": "79bf6545-3395-4966-9ef3-90dc42b3e780", 00:22:58.065 "strip_size_kb": 0, 00:22:58.065 "state": "online", 00:22:58.065 "raid_level": "raid1", 00:22:58.065 "superblock": false, 00:22:58.065 "num_base_bdevs": 3, 00:22:58.065 "num_base_bdevs_discovered": 3, 00:22:58.065 "num_base_bdevs_operational": 3, 00:22:58.065 "base_bdevs_list": [ 00:22:58.065 { 00:22:58.065 "name": "NewBaseBdev", 00:22:58.065 "uuid": "c4b109f3-8702-4b41-88d5-1ea0b7448817", 00:22:58.065 "is_configured": true, 00:22:58.065 "data_offset": 0, 00:22:58.065 "data_size": 65536 00:22:58.065 }, 00:22:58.065 { 00:22:58.065 "name": "BaseBdev2", 00:22:58.065 "uuid": "2491c869-b47f-4966-9f63-31131e05b1c3", 00:22:58.065 "is_configured": true, 00:22:58.065 "data_offset": 0, 00:22:58.065 "data_size": 65536 00:22:58.065 }, 00:22:58.065 { 00:22:58.065 "name": "BaseBdev3", 00:22:58.065 "uuid": "e90ddc37-b7ef-4556-8097-36986a3a88e3", 00:22:58.065 "is_configured": true, 00:22:58.065 "data_offset": 0, 00:22:58.065 "data_size": 65536 00:22:58.065 } 00:22:58.065 ] 00:22:58.065 } 00:22:58.065 } 00:22:58.065 }' 00:22:58.065 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:22:58.065 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:22:58.065 BaseBdev2 00:22:58.065 BaseBdev3' 00:22:58.065 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:22:58.324 [2024-10-01 20:21:53.514029] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:22:58.324 [2024-10-01 20:21:53.514282] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:22:58.324 [2024-10-01 20:21:53.514419] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:22:58.324 [2024-10-01 20:21:53.514864] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:22:58.324 [2024-10-01 20:21:53.514886] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 68036 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 68036 ']' 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 68036 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 68036 00:22:58.324 killing process with pid 68036 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 68036' 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 68036 00:22:58.324 [2024-10-01 20:21:53.558208] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:22:58.324 20:21:53 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 68036 00:22:58.891 [2024-10-01 20:21:53.837804] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:23:00.791 00:23:00.791 real 0m12.812s 00:23:00.791 user 0m20.546s 00:23:00.791 sys 0m1.820s 00:23:00.791 ************************************ 00:23:00.791 END TEST raid_state_function_test 00:23:00.791 ************************************ 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:00.791 20:21:55 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 3 true 00:23:00.791 20:21:55 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:23:00.791 20:21:55 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:23:00.791 20:21:55 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:23:00.791 ************************************ 00:23:00.791 START TEST raid_state_function_test_sb 00:23:00.791 ************************************ 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 3 true 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:23:00.791 Process raid pid: 68683 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=68683 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 68683' 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 68683 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 68683 ']' 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:00.791 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:00.791 20:21:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:00.791 [2024-10-01 20:21:55.826490] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:23:00.791 [2024-10-01 20:21:55.826919] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:00.791 [2024-10-01 20:21:55.998867] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:01.416 [2024-10-01 20:21:56.314102] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:23:01.416 [2024-10-01 20:21:56.536970] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:23:01.416 [2024-10-01 20:21:56.537311] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:01.993 [2024-10-01 20:21:56.973490] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:23:01.993 [2024-10-01 20:21:56.973740] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:23:01.993 [2024-10-01 20:21:56.973891] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:23:01.993 [2024-10-01 20:21:56.973966] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:23:01.993 [2024-10-01 20:21:56.974102] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:23:01.993 [2024-10-01 20:21:56.974176] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:01.993 20:21:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:01.993 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:01.993 "name": "Existed_Raid", 00:23:01.993 "uuid": "048b488d-0a49-47e3-a3b9-4ec71c37be18", 00:23:01.993 "strip_size_kb": 0, 00:23:01.993 "state": "configuring", 00:23:01.993 "raid_level": "raid1", 00:23:01.993 "superblock": true, 00:23:01.993 "num_base_bdevs": 3, 00:23:01.993 "num_base_bdevs_discovered": 0, 00:23:01.993 "num_base_bdevs_operational": 3, 00:23:01.993 "base_bdevs_list": [ 00:23:01.993 { 00:23:01.993 "name": "BaseBdev1", 00:23:01.993 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:01.993 "is_configured": false, 00:23:01.993 "data_offset": 0, 00:23:01.993 "data_size": 0 00:23:01.993 }, 00:23:01.993 { 00:23:01.993 "name": "BaseBdev2", 00:23:01.993 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:01.993 "is_configured": false, 00:23:01.993 "data_offset": 0, 00:23:01.993 "data_size": 0 00:23:01.993 }, 00:23:01.993 { 00:23:01.993 "name": "BaseBdev3", 00:23:01.993 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:01.993 "is_configured": false, 00:23:01.993 "data_offset": 0, 00:23:01.993 "data_size": 0 00:23:01.993 } 00:23:01.993 ] 00:23:01.993 }' 00:23:01.993 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:01.993 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:02.251 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:23:02.251 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:02.251 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:02.510 [2024-10-01 20:21:57.505515] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:23:02.510 [2024-10-01 20:21:57.505575] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:02.510 [2024-10-01 20:21:57.513512] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:23:02.510 [2024-10-01 20:21:57.513748] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:23:02.510 [2024-10-01 20:21:57.513928] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:23:02.510 [2024-10-01 20:21:57.514007] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:23:02.510 [2024-10-01 20:21:57.514129] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:23:02.510 [2024-10-01 20:21:57.514201] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:02.510 [2024-10-01 20:21:57.564189] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:23:02.510 BaseBdev1 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:02.510 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:02.510 [ 00:23:02.510 { 00:23:02.510 "name": "BaseBdev1", 00:23:02.510 "aliases": [ 00:23:02.510 "a656a866-0808-4b92-8ef3-f0afe621ea87" 00:23:02.510 ], 00:23:02.510 "product_name": "Malloc disk", 00:23:02.510 "block_size": 512, 00:23:02.510 "num_blocks": 65536, 00:23:02.510 "uuid": "a656a866-0808-4b92-8ef3-f0afe621ea87", 00:23:02.510 "assigned_rate_limits": { 00:23:02.510 "rw_ios_per_sec": 0, 00:23:02.510 "rw_mbytes_per_sec": 0, 00:23:02.510 "r_mbytes_per_sec": 0, 00:23:02.510 "w_mbytes_per_sec": 0 00:23:02.510 }, 00:23:02.510 "claimed": true, 00:23:02.510 "claim_type": "exclusive_write", 00:23:02.510 "zoned": false, 00:23:02.510 "supported_io_types": { 00:23:02.510 "read": true, 00:23:02.510 "write": true, 00:23:02.510 "unmap": true, 00:23:02.510 "flush": true, 00:23:02.510 "reset": true, 00:23:02.510 "nvme_admin": false, 00:23:02.510 "nvme_io": false, 00:23:02.510 "nvme_io_md": false, 00:23:02.510 "write_zeroes": true, 00:23:02.510 "zcopy": true, 00:23:02.510 "get_zone_info": false, 00:23:02.510 "zone_management": false, 00:23:02.510 "zone_append": false, 00:23:02.510 "compare": false, 00:23:02.510 "compare_and_write": false, 00:23:02.510 "abort": true, 00:23:02.510 "seek_hole": false, 00:23:02.510 "seek_data": false, 00:23:02.510 "copy": true, 00:23:02.510 "nvme_iov_md": false 00:23:02.510 }, 00:23:02.510 "memory_domains": [ 00:23:02.511 { 00:23:02.511 "dma_device_id": "system", 00:23:02.511 "dma_device_type": 1 00:23:02.511 }, 00:23:02.511 { 00:23:02.511 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:02.511 "dma_device_type": 2 00:23:02.511 } 00:23:02.511 ], 00:23:02.511 "driver_specific": {} 00:23:02.511 } 00:23:02.511 ] 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:02.511 "name": "Existed_Raid", 00:23:02.511 "uuid": "83688e07-ac49-4771-ac71-19e8678f9ef6", 00:23:02.511 "strip_size_kb": 0, 00:23:02.511 "state": "configuring", 00:23:02.511 "raid_level": "raid1", 00:23:02.511 "superblock": true, 00:23:02.511 "num_base_bdevs": 3, 00:23:02.511 "num_base_bdevs_discovered": 1, 00:23:02.511 "num_base_bdevs_operational": 3, 00:23:02.511 "base_bdevs_list": [ 00:23:02.511 { 00:23:02.511 "name": "BaseBdev1", 00:23:02.511 "uuid": "a656a866-0808-4b92-8ef3-f0afe621ea87", 00:23:02.511 "is_configured": true, 00:23:02.511 "data_offset": 2048, 00:23:02.511 "data_size": 63488 00:23:02.511 }, 00:23:02.511 { 00:23:02.511 "name": "BaseBdev2", 00:23:02.511 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:02.511 "is_configured": false, 00:23:02.511 "data_offset": 0, 00:23:02.511 "data_size": 0 00:23:02.511 }, 00:23:02.511 { 00:23:02.511 "name": "BaseBdev3", 00:23:02.511 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:02.511 "is_configured": false, 00:23:02.511 "data_offset": 0, 00:23:02.511 "data_size": 0 00:23:02.511 } 00:23:02.511 ] 00:23:02.511 }' 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:02.511 20:21:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:03.078 [2024-10-01 20:21:58.104423] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:23:03.078 [2024-10-01 20:21:58.104502] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:03.078 [2024-10-01 20:21:58.112443] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:23:03.078 [2024-10-01 20:21:58.115284] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:23:03.078 [2024-10-01 20:21:58.115483] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:23:03.078 [2024-10-01 20:21:58.115618] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:23:03.078 [2024-10-01 20:21:58.115692] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:03.078 "name": "Existed_Raid", 00:23:03.078 "uuid": "c3ad3613-b173-41dd-be9f-a78f585adf89", 00:23:03.078 "strip_size_kb": 0, 00:23:03.078 "state": "configuring", 00:23:03.078 "raid_level": "raid1", 00:23:03.078 "superblock": true, 00:23:03.078 "num_base_bdevs": 3, 00:23:03.078 "num_base_bdevs_discovered": 1, 00:23:03.078 "num_base_bdevs_operational": 3, 00:23:03.078 "base_bdevs_list": [ 00:23:03.078 { 00:23:03.078 "name": "BaseBdev1", 00:23:03.078 "uuid": "a656a866-0808-4b92-8ef3-f0afe621ea87", 00:23:03.078 "is_configured": true, 00:23:03.078 "data_offset": 2048, 00:23:03.078 "data_size": 63488 00:23:03.078 }, 00:23:03.078 { 00:23:03.078 "name": "BaseBdev2", 00:23:03.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:03.078 "is_configured": false, 00:23:03.078 "data_offset": 0, 00:23:03.078 "data_size": 0 00:23:03.078 }, 00:23:03.078 { 00:23:03.078 "name": "BaseBdev3", 00:23:03.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:03.078 "is_configured": false, 00:23:03.078 "data_offset": 0, 00:23:03.078 "data_size": 0 00:23:03.078 } 00:23:03.078 ] 00:23:03.078 }' 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:03.078 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:03.645 [2024-10-01 20:21:58.683666] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:23:03.645 BaseBdev2 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:03.645 [ 00:23:03.645 { 00:23:03.645 "name": "BaseBdev2", 00:23:03.645 "aliases": [ 00:23:03.645 "4177ba56-46cf-4b76-bdc5-f47784a66e23" 00:23:03.645 ], 00:23:03.645 "product_name": "Malloc disk", 00:23:03.645 "block_size": 512, 00:23:03.645 "num_blocks": 65536, 00:23:03.645 "uuid": "4177ba56-46cf-4b76-bdc5-f47784a66e23", 00:23:03.645 "assigned_rate_limits": { 00:23:03.645 "rw_ios_per_sec": 0, 00:23:03.645 "rw_mbytes_per_sec": 0, 00:23:03.645 "r_mbytes_per_sec": 0, 00:23:03.645 "w_mbytes_per_sec": 0 00:23:03.645 }, 00:23:03.645 "claimed": true, 00:23:03.645 "claim_type": "exclusive_write", 00:23:03.645 "zoned": false, 00:23:03.645 "supported_io_types": { 00:23:03.645 "read": true, 00:23:03.645 "write": true, 00:23:03.645 "unmap": true, 00:23:03.645 "flush": true, 00:23:03.645 "reset": true, 00:23:03.645 "nvme_admin": false, 00:23:03.645 "nvme_io": false, 00:23:03.645 "nvme_io_md": false, 00:23:03.645 "write_zeroes": true, 00:23:03.645 "zcopy": true, 00:23:03.645 "get_zone_info": false, 00:23:03.645 "zone_management": false, 00:23:03.645 "zone_append": false, 00:23:03.645 "compare": false, 00:23:03.645 "compare_and_write": false, 00:23:03.645 "abort": true, 00:23:03.645 "seek_hole": false, 00:23:03.645 "seek_data": false, 00:23:03.645 "copy": true, 00:23:03.645 "nvme_iov_md": false 00:23:03.645 }, 00:23:03.645 "memory_domains": [ 00:23:03.645 { 00:23:03.645 "dma_device_id": "system", 00:23:03.645 "dma_device_type": 1 00:23:03.645 }, 00:23:03.645 { 00:23:03.645 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:03.645 "dma_device_type": 2 00:23:03.645 } 00:23:03.645 ], 00:23:03.645 "driver_specific": {} 00:23:03.645 } 00:23:03.645 ] 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:03.645 "name": "Existed_Raid", 00:23:03.645 "uuid": "c3ad3613-b173-41dd-be9f-a78f585adf89", 00:23:03.645 "strip_size_kb": 0, 00:23:03.645 "state": "configuring", 00:23:03.645 "raid_level": "raid1", 00:23:03.645 "superblock": true, 00:23:03.645 "num_base_bdevs": 3, 00:23:03.645 "num_base_bdevs_discovered": 2, 00:23:03.645 "num_base_bdevs_operational": 3, 00:23:03.645 "base_bdevs_list": [ 00:23:03.645 { 00:23:03.645 "name": "BaseBdev1", 00:23:03.645 "uuid": "a656a866-0808-4b92-8ef3-f0afe621ea87", 00:23:03.645 "is_configured": true, 00:23:03.645 "data_offset": 2048, 00:23:03.645 "data_size": 63488 00:23:03.645 }, 00:23:03.645 { 00:23:03.645 "name": "BaseBdev2", 00:23:03.645 "uuid": "4177ba56-46cf-4b76-bdc5-f47784a66e23", 00:23:03.645 "is_configured": true, 00:23:03.645 "data_offset": 2048, 00:23:03.645 "data_size": 63488 00:23:03.645 }, 00:23:03.645 { 00:23:03.645 "name": "BaseBdev3", 00:23:03.645 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:03.645 "is_configured": false, 00:23:03.645 "data_offset": 0, 00:23:03.645 "data_size": 0 00:23:03.645 } 00:23:03.645 ] 00:23:03.645 }' 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:03.645 20:21:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:04.211 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:23:04.211 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:04.211 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:04.211 [2024-10-01 20:21:59.318971] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:23:04.211 [2024-10-01 20:21:59.319349] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:23:04.211 [2024-10-01 20:21:59.319389] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:23:04.211 [2024-10-01 20:21:59.319768] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:23:04.211 BaseBdev3 00:23:04.211 [2024-10-01 20:21:59.319982] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:23:04.211 [2024-10-01 20:21:59.320002] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:23:04.211 [2024-10-01 20:21:59.320203] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:04.211 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:04.211 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:23:04.211 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:23:04.211 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:04.211 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:23:04.211 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:04.211 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:04.211 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:04.211 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:04.212 [ 00:23:04.212 { 00:23:04.212 "name": "BaseBdev3", 00:23:04.212 "aliases": [ 00:23:04.212 "dabeb9df-317e-479f-b8aa-ee2fd785e488" 00:23:04.212 ], 00:23:04.212 "product_name": "Malloc disk", 00:23:04.212 "block_size": 512, 00:23:04.212 "num_blocks": 65536, 00:23:04.212 "uuid": "dabeb9df-317e-479f-b8aa-ee2fd785e488", 00:23:04.212 "assigned_rate_limits": { 00:23:04.212 "rw_ios_per_sec": 0, 00:23:04.212 "rw_mbytes_per_sec": 0, 00:23:04.212 "r_mbytes_per_sec": 0, 00:23:04.212 "w_mbytes_per_sec": 0 00:23:04.212 }, 00:23:04.212 "claimed": true, 00:23:04.212 "claim_type": "exclusive_write", 00:23:04.212 "zoned": false, 00:23:04.212 "supported_io_types": { 00:23:04.212 "read": true, 00:23:04.212 "write": true, 00:23:04.212 "unmap": true, 00:23:04.212 "flush": true, 00:23:04.212 "reset": true, 00:23:04.212 "nvme_admin": false, 00:23:04.212 "nvme_io": false, 00:23:04.212 "nvme_io_md": false, 00:23:04.212 "write_zeroes": true, 00:23:04.212 "zcopy": true, 00:23:04.212 "get_zone_info": false, 00:23:04.212 "zone_management": false, 00:23:04.212 "zone_append": false, 00:23:04.212 "compare": false, 00:23:04.212 "compare_and_write": false, 00:23:04.212 "abort": true, 00:23:04.212 "seek_hole": false, 00:23:04.212 "seek_data": false, 00:23:04.212 "copy": true, 00:23:04.212 "nvme_iov_md": false 00:23:04.212 }, 00:23:04.212 "memory_domains": [ 00:23:04.212 { 00:23:04.212 "dma_device_id": "system", 00:23:04.212 "dma_device_type": 1 00:23:04.212 }, 00:23:04.212 { 00:23:04.212 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:04.212 "dma_device_type": 2 00:23:04.212 } 00:23:04.212 ], 00:23:04.212 "driver_specific": {} 00:23:04.212 } 00:23:04.212 ] 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:04.212 "name": "Existed_Raid", 00:23:04.212 "uuid": "c3ad3613-b173-41dd-be9f-a78f585adf89", 00:23:04.212 "strip_size_kb": 0, 00:23:04.212 "state": "online", 00:23:04.212 "raid_level": "raid1", 00:23:04.212 "superblock": true, 00:23:04.212 "num_base_bdevs": 3, 00:23:04.212 "num_base_bdevs_discovered": 3, 00:23:04.212 "num_base_bdevs_operational": 3, 00:23:04.212 "base_bdevs_list": [ 00:23:04.212 { 00:23:04.212 "name": "BaseBdev1", 00:23:04.212 "uuid": "a656a866-0808-4b92-8ef3-f0afe621ea87", 00:23:04.212 "is_configured": true, 00:23:04.212 "data_offset": 2048, 00:23:04.212 "data_size": 63488 00:23:04.212 }, 00:23:04.212 { 00:23:04.212 "name": "BaseBdev2", 00:23:04.212 "uuid": "4177ba56-46cf-4b76-bdc5-f47784a66e23", 00:23:04.212 "is_configured": true, 00:23:04.212 "data_offset": 2048, 00:23:04.212 "data_size": 63488 00:23:04.212 }, 00:23:04.212 { 00:23:04.212 "name": "BaseBdev3", 00:23:04.212 "uuid": "dabeb9df-317e-479f-b8aa-ee2fd785e488", 00:23:04.212 "is_configured": true, 00:23:04.212 "data_offset": 2048, 00:23:04.212 "data_size": 63488 00:23:04.212 } 00:23:04.212 ] 00:23:04.212 }' 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:04.212 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:04.777 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:23:04.777 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:23:04.777 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:23:04.778 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:23:04.778 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:23:04.778 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:23:04.778 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:23:04.778 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:23:04.778 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:04.778 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:04.778 [2024-10-01 20:21:59.895644] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:23:04.778 20:21:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:04.778 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:04.778 "name": "Existed_Raid", 00:23:04.778 "aliases": [ 00:23:04.778 "c3ad3613-b173-41dd-be9f-a78f585adf89" 00:23:04.778 ], 00:23:04.778 "product_name": "Raid Volume", 00:23:04.778 "block_size": 512, 00:23:04.778 "num_blocks": 63488, 00:23:04.778 "uuid": "c3ad3613-b173-41dd-be9f-a78f585adf89", 00:23:04.778 "assigned_rate_limits": { 00:23:04.778 "rw_ios_per_sec": 0, 00:23:04.778 "rw_mbytes_per_sec": 0, 00:23:04.778 "r_mbytes_per_sec": 0, 00:23:04.778 "w_mbytes_per_sec": 0 00:23:04.778 }, 00:23:04.778 "claimed": false, 00:23:04.778 "zoned": false, 00:23:04.778 "supported_io_types": { 00:23:04.778 "read": true, 00:23:04.778 "write": true, 00:23:04.778 "unmap": false, 00:23:04.778 "flush": false, 00:23:04.778 "reset": true, 00:23:04.778 "nvme_admin": false, 00:23:04.778 "nvme_io": false, 00:23:04.778 "nvme_io_md": false, 00:23:04.778 "write_zeroes": true, 00:23:04.778 "zcopy": false, 00:23:04.778 "get_zone_info": false, 00:23:04.778 "zone_management": false, 00:23:04.778 "zone_append": false, 00:23:04.778 "compare": false, 00:23:04.778 "compare_and_write": false, 00:23:04.778 "abort": false, 00:23:04.778 "seek_hole": false, 00:23:04.778 "seek_data": false, 00:23:04.778 "copy": false, 00:23:04.778 "nvme_iov_md": false 00:23:04.778 }, 00:23:04.778 "memory_domains": [ 00:23:04.778 { 00:23:04.778 "dma_device_id": "system", 00:23:04.778 "dma_device_type": 1 00:23:04.778 }, 00:23:04.778 { 00:23:04.778 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:04.778 "dma_device_type": 2 00:23:04.778 }, 00:23:04.778 { 00:23:04.778 "dma_device_id": "system", 00:23:04.778 "dma_device_type": 1 00:23:04.778 }, 00:23:04.778 { 00:23:04.778 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:04.778 "dma_device_type": 2 00:23:04.778 }, 00:23:04.778 { 00:23:04.778 "dma_device_id": "system", 00:23:04.778 "dma_device_type": 1 00:23:04.778 }, 00:23:04.778 { 00:23:04.778 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:04.778 "dma_device_type": 2 00:23:04.778 } 00:23:04.778 ], 00:23:04.778 "driver_specific": { 00:23:04.778 "raid": { 00:23:04.778 "uuid": "c3ad3613-b173-41dd-be9f-a78f585adf89", 00:23:04.778 "strip_size_kb": 0, 00:23:04.778 "state": "online", 00:23:04.778 "raid_level": "raid1", 00:23:04.778 "superblock": true, 00:23:04.778 "num_base_bdevs": 3, 00:23:04.778 "num_base_bdevs_discovered": 3, 00:23:04.778 "num_base_bdevs_operational": 3, 00:23:04.778 "base_bdevs_list": [ 00:23:04.778 { 00:23:04.778 "name": "BaseBdev1", 00:23:04.778 "uuid": "a656a866-0808-4b92-8ef3-f0afe621ea87", 00:23:04.778 "is_configured": true, 00:23:04.778 "data_offset": 2048, 00:23:04.778 "data_size": 63488 00:23:04.778 }, 00:23:04.778 { 00:23:04.778 "name": "BaseBdev2", 00:23:04.778 "uuid": "4177ba56-46cf-4b76-bdc5-f47784a66e23", 00:23:04.778 "is_configured": true, 00:23:04.778 "data_offset": 2048, 00:23:04.778 "data_size": 63488 00:23:04.778 }, 00:23:04.778 { 00:23:04.778 "name": "BaseBdev3", 00:23:04.778 "uuid": "dabeb9df-317e-479f-b8aa-ee2fd785e488", 00:23:04.778 "is_configured": true, 00:23:04.778 "data_offset": 2048, 00:23:04.778 "data_size": 63488 00:23:04.778 } 00:23:04.778 ] 00:23:04.778 } 00:23:04.778 } 00:23:04.778 }' 00:23:04.778 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:23:04.778 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:23:04.778 BaseBdev2 00:23:04.778 BaseBdev3' 00:23:04.778 20:21:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:05.036 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:05.036 [2024-10-01 20:22:00.239350] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:05.298 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:05.299 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:05.299 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:05.299 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:05.299 "name": "Existed_Raid", 00:23:05.299 "uuid": "c3ad3613-b173-41dd-be9f-a78f585adf89", 00:23:05.299 "strip_size_kb": 0, 00:23:05.299 "state": "online", 00:23:05.299 "raid_level": "raid1", 00:23:05.299 "superblock": true, 00:23:05.299 "num_base_bdevs": 3, 00:23:05.299 "num_base_bdevs_discovered": 2, 00:23:05.299 "num_base_bdevs_operational": 2, 00:23:05.299 "base_bdevs_list": [ 00:23:05.299 { 00:23:05.299 "name": null, 00:23:05.299 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:05.299 "is_configured": false, 00:23:05.299 "data_offset": 0, 00:23:05.299 "data_size": 63488 00:23:05.299 }, 00:23:05.299 { 00:23:05.299 "name": "BaseBdev2", 00:23:05.299 "uuid": "4177ba56-46cf-4b76-bdc5-f47784a66e23", 00:23:05.299 "is_configured": true, 00:23:05.299 "data_offset": 2048, 00:23:05.299 "data_size": 63488 00:23:05.299 }, 00:23:05.299 { 00:23:05.299 "name": "BaseBdev3", 00:23:05.299 "uuid": "dabeb9df-317e-479f-b8aa-ee2fd785e488", 00:23:05.299 "is_configured": true, 00:23:05.299 "data_offset": 2048, 00:23:05.299 "data_size": 63488 00:23:05.299 } 00:23:05.299 ] 00:23:05.299 }' 00:23:05.299 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:05.299 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:05.864 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:23:05.864 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:23:05.864 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:05.864 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:05.864 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:05.864 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:23:05.864 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:05.865 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:23:05.865 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:23:05.865 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:23:05.865 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:05.865 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:05.865 [2024-10-01 20:22:00.882487] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:23:05.865 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:05.865 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:23:05.865 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:23:05.865 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:05.865 20:22:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:23:05.865 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:05.865 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:05.865 20:22:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:05.865 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:23:05.865 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:23:05.865 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:23:05.865 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:05.865 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:05.865 [2024-10-01 20:22:01.029767] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:23:05.865 [2024-10-01 20:22:01.030089] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:23:06.123 [2024-10-01 20:22:01.118815] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:23:06.123 [2024-10-01 20:22:01.119166] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:23:06.123 [2024-10-01 20:22:01.119208] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:23:06.123 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:06.123 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:23:06.123 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:23:06.123 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:06.123 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:23:06.123 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:06.123 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:06.123 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:06.123 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:23:06.123 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:23:06.123 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:23:06.123 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:06.124 BaseBdev2 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:06.124 [ 00:23:06.124 { 00:23:06.124 "name": "BaseBdev2", 00:23:06.124 "aliases": [ 00:23:06.124 "c93c52bb-c0ce-455c-9bca-bce75d2807ff" 00:23:06.124 ], 00:23:06.124 "product_name": "Malloc disk", 00:23:06.124 "block_size": 512, 00:23:06.124 "num_blocks": 65536, 00:23:06.124 "uuid": "c93c52bb-c0ce-455c-9bca-bce75d2807ff", 00:23:06.124 "assigned_rate_limits": { 00:23:06.124 "rw_ios_per_sec": 0, 00:23:06.124 "rw_mbytes_per_sec": 0, 00:23:06.124 "r_mbytes_per_sec": 0, 00:23:06.124 "w_mbytes_per_sec": 0 00:23:06.124 }, 00:23:06.124 "claimed": false, 00:23:06.124 "zoned": false, 00:23:06.124 "supported_io_types": { 00:23:06.124 "read": true, 00:23:06.124 "write": true, 00:23:06.124 "unmap": true, 00:23:06.124 "flush": true, 00:23:06.124 "reset": true, 00:23:06.124 "nvme_admin": false, 00:23:06.124 "nvme_io": false, 00:23:06.124 "nvme_io_md": false, 00:23:06.124 "write_zeroes": true, 00:23:06.124 "zcopy": true, 00:23:06.124 "get_zone_info": false, 00:23:06.124 "zone_management": false, 00:23:06.124 "zone_append": false, 00:23:06.124 "compare": false, 00:23:06.124 "compare_and_write": false, 00:23:06.124 "abort": true, 00:23:06.124 "seek_hole": false, 00:23:06.124 "seek_data": false, 00:23:06.124 "copy": true, 00:23:06.124 "nvme_iov_md": false 00:23:06.124 }, 00:23:06.124 "memory_domains": [ 00:23:06.124 { 00:23:06.124 "dma_device_id": "system", 00:23:06.124 "dma_device_type": 1 00:23:06.124 }, 00:23:06.124 { 00:23:06.124 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:06.124 "dma_device_type": 2 00:23:06.124 } 00:23:06.124 ], 00:23:06.124 "driver_specific": {} 00:23:06.124 } 00:23:06.124 ] 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:06.124 BaseBdev3 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:06.124 [ 00:23:06.124 { 00:23:06.124 "name": "BaseBdev3", 00:23:06.124 "aliases": [ 00:23:06.124 "df971be6-3f78-4484-b8e6-d8807b482b40" 00:23:06.124 ], 00:23:06.124 "product_name": "Malloc disk", 00:23:06.124 "block_size": 512, 00:23:06.124 "num_blocks": 65536, 00:23:06.124 "uuid": "df971be6-3f78-4484-b8e6-d8807b482b40", 00:23:06.124 "assigned_rate_limits": { 00:23:06.124 "rw_ios_per_sec": 0, 00:23:06.124 "rw_mbytes_per_sec": 0, 00:23:06.124 "r_mbytes_per_sec": 0, 00:23:06.124 "w_mbytes_per_sec": 0 00:23:06.124 }, 00:23:06.124 "claimed": false, 00:23:06.124 "zoned": false, 00:23:06.124 "supported_io_types": { 00:23:06.124 "read": true, 00:23:06.124 "write": true, 00:23:06.124 "unmap": true, 00:23:06.124 "flush": true, 00:23:06.124 "reset": true, 00:23:06.124 "nvme_admin": false, 00:23:06.124 "nvme_io": false, 00:23:06.124 "nvme_io_md": false, 00:23:06.124 "write_zeroes": true, 00:23:06.124 "zcopy": true, 00:23:06.124 "get_zone_info": false, 00:23:06.124 "zone_management": false, 00:23:06.124 "zone_append": false, 00:23:06.124 "compare": false, 00:23:06.124 "compare_and_write": false, 00:23:06.124 "abort": true, 00:23:06.124 "seek_hole": false, 00:23:06.124 "seek_data": false, 00:23:06.124 "copy": true, 00:23:06.124 "nvme_iov_md": false 00:23:06.124 }, 00:23:06.124 "memory_domains": [ 00:23:06.124 { 00:23:06.124 "dma_device_id": "system", 00:23:06.124 "dma_device_type": 1 00:23:06.124 }, 00:23:06.124 { 00:23:06.124 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:06.124 "dma_device_type": 2 00:23:06.124 } 00:23:06.124 ], 00:23:06.124 "driver_specific": {} 00:23:06.124 } 00:23:06.124 ] 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:06.124 [2024-10-01 20:22:01.332071] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:23:06.124 [2024-10-01 20:22:01.332153] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:23:06.124 [2024-10-01 20:22:01.332204] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:23:06.124 [2024-10-01 20:22:01.335477] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:06.124 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:06.125 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:06.125 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:06.125 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:06.125 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:06.125 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:06.125 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:06.125 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:06.125 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:06.383 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:06.383 "name": "Existed_Raid", 00:23:06.383 "uuid": "841c252b-cbdb-44df-b0d8-82ac6b43feb0", 00:23:06.383 "strip_size_kb": 0, 00:23:06.383 "state": "configuring", 00:23:06.383 "raid_level": "raid1", 00:23:06.383 "superblock": true, 00:23:06.383 "num_base_bdevs": 3, 00:23:06.383 "num_base_bdevs_discovered": 2, 00:23:06.383 "num_base_bdevs_operational": 3, 00:23:06.383 "base_bdevs_list": [ 00:23:06.383 { 00:23:06.383 "name": "BaseBdev1", 00:23:06.383 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:06.383 "is_configured": false, 00:23:06.383 "data_offset": 0, 00:23:06.383 "data_size": 0 00:23:06.383 }, 00:23:06.383 { 00:23:06.383 "name": "BaseBdev2", 00:23:06.383 "uuid": "c93c52bb-c0ce-455c-9bca-bce75d2807ff", 00:23:06.383 "is_configured": true, 00:23:06.383 "data_offset": 2048, 00:23:06.383 "data_size": 63488 00:23:06.383 }, 00:23:06.383 { 00:23:06.383 "name": "BaseBdev3", 00:23:06.383 "uuid": "df971be6-3f78-4484-b8e6-d8807b482b40", 00:23:06.383 "is_configured": true, 00:23:06.383 "data_offset": 2048, 00:23:06.383 "data_size": 63488 00:23:06.383 } 00:23:06.383 ] 00:23:06.383 }' 00:23:06.383 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:06.383 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:06.642 [2024-10-01 20:22:01.836162] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:06.642 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:06.642 "name": "Existed_Raid", 00:23:06.642 "uuid": "841c252b-cbdb-44df-b0d8-82ac6b43feb0", 00:23:06.642 "strip_size_kb": 0, 00:23:06.642 "state": "configuring", 00:23:06.642 "raid_level": "raid1", 00:23:06.642 "superblock": true, 00:23:06.642 "num_base_bdevs": 3, 00:23:06.642 "num_base_bdevs_discovered": 1, 00:23:06.642 "num_base_bdevs_operational": 3, 00:23:06.642 "base_bdevs_list": [ 00:23:06.642 { 00:23:06.642 "name": "BaseBdev1", 00:23:06.642 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:06.642 "is_configured": false, 00:23:06.642 "data_offset": 0, 00:23:06.642 "data_size": 0 00:23:06.643 }, 00:23:06.643 { 00:23:06.643 "name": null, 00:23:06.643 "uuid": "c93c52bb-c0ce-455c-9bca-bce75d2807ff", 00:23:06.643 "is_configured": false, 00:23:06.643 "data_offset": 0, 00:23:06.643 "data_size": 63488 00:23:06.643 }, 00:23:06.643 { 00:23:06.643 "name": "BaseBdev3", 00:23:06.643 "uuid": "df971be6-3f78-4484-b8e6-d8807b482b40", 00:23:06.643 "is_configured": true, 00:23:06.643 "data_offset": 2048, 00:23:06.643 "data_size": 63488 00:23:06.643 } 00:23:06.643 ] 00:23:06.643 }' 00:23:06.643 20:22:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:06.643 20:22:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:07.239 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:07.239 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:07.239 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:07.239 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:23:07.239 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:07.239 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:23:07.239 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:23:07.239 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:07.239 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:07.239 [2024-10-01 20:22:02.431849] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:23:07.239 BaseBdev1 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:07.240 [ 00:23:07.240 { 00:23:07.240 "name": "BaseBdev1", 00:23:07.240 "aliases": [ 00:23:07.240 "2b9ea74d-94fb-4acc-b9ba-aa13e51191ff" 00:23:07.240 ], 00:23:07.240 "product_name": "Malloc disk", 00:23:07.240 "block_size": 512, 00:23:07.240 "num_blocks": 65536, 00:23:07.240 "uuid": "2b9ea74d-94fb-4acc-b9ba-aa13e51191ff", 00:23:07.240 "assigned_rate_limits": { 00:23:07.240 "rw_ios_per_sec": 0, 00:23:07.240 "rw_mbytes_per_sec": 0, 00:23:07.240 "r_mbytes_per_sec": 0, 00:23:07.240 "w_mbytes_per_sec": 0 00:23:07.240 }, 00:23:07.240 "claimed": true, 00:23:07.240 "claim_type": "exclusive_write", 00:23:07.240 "zoned": false, 00:23:07.240 "supported_io_types": { 00:23:07.240 "read": true, 00:23:07.240 "write": true, 00:23:07.240 "unmap": true, 00:23:07.240 "flush": true, 00:23:07.240 "reset": true, 00:23:07.240 "nvme_admin": false, 00:23:07.240 "nvme_io": false, 00:23:07.240 "nvme_io_md": false, 00:23:07.240 "write_zeroes": true, 00:23:07.240 "zcopy": true, 00:23:07.240 "get_zone_info": false, 00:23:07.240 "zone_management": false, 00:23:07.240 "zone_append": false, 00:23:07.240 "compare": false, 00:23:07.240 "compare_and_write": false, 00:23:07.240 "abort": true, 00:23:07.240 "seek_hole": false, 00:23:07.240 "seek_data": false, 00:23:07.240 "copy": true, 00:23:07.240 "nvme_iov_md": false 00:23:07.240 }, 00:23:07.240 "memory_domains": [ 00:23:07.240 { 00:23:07.240 "dma_device_id": "system", 00:23:07.240 "dma_device_type": 1 00:23:07.240 }, 00:23:07.240 { 00:23:07.240 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:07.240 "dma_device_type": 2 00:23:07.240 } 00:23:07.240 ], 00:23:07.240 "driver_specific": {} 00:23:07.240 } 00:23:07.240 ] 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:07.240 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:07.498 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:07.498 "name": "Existed_Raid", 00:23:07.498 "uuid": "841c252b-cbdb-44df-b0d8-82ac6b43feb0", 00:23:07.498 "strip_size_kb": 0, 00:23:07.498 "state": "configuring", 00:23:07.498 "raid_level": "raid1", 00:23:07.498 "superblock": true, 00:23:07.498 "num_base_bdevs": 3, 00:23:07.498 "num_base_bdevs_discovered": 2, 00:23:07.498 "num_base_bdevs_operational": 3, 00:23:07.498 "base_bdevs_list": [ 00:23:07.498 { 00:23:07.498 "name": "BaseBdev1", 00:23:07.498 "uuid": "2b9ea74d-94fb-4acc-b9ba-aa13e51191ff", 00:23:07.498 "is_configured": true, 00:23:07.498 "data_offset": 2048, 00:23:07.498 "data_size": 63488 00:23:07.498 }, 00:23:07.498 { 00:23:07.498 "name": null, 00:23:07.498 "uuid": "c93c52bb-c0ce-455c-9bca-bce75d2807ff", 00:23:07.498 "is_configured": false, 00:23:07.498 "data_offset": 0, 00:23:07.498 "data_size": 63488 00:23:07.498 }, 00:23:07.498 { 00:23:07.498 "name": "BaseBdev3", 00:23:07.498 "uuid": "df971be6-3f78-4484-b8e6-d8807b482b40", 00:23:07.498 "is_configured": true, 00:23:07.498 "data_offset": 2048, 00:23:07.498 "data_size": 63488 00:23:07.498 } 00:23:07.498 ] 00:23:07.498 }' 00:23:07.498 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:07.498 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:07.757 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:07.757 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:07.757 20:22:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:23:07.757 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:07.757 20:22:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:08.017 [2024-10-01 20:22:03.040114] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:08.017 "name": "Existed_Raid", 00:23:08.017 "uuid": "841c252b-cbdb-44df-b0d8-82ac6b43feb0", 00:23:08.017 "strip_size_kb": 0, 00:23:08.017 "state": "configuring", 00:23:08.017 "raid_level": "raid1", 00:23:08.017 "superblock": true, 00:23:08.017 "num_base_bdevs": 3, 00:23:08.017 "num_base_bdevs_discovered": 1, 00:23:08.017 "num_base_bdevs_operational": 3, 00:23:08.017 "base_bdevs_list": [ 00:23:08.017 { 00:23:08.017 "name": "BaseBdev1", 00:23:08.017 "uuid": "2b9ea74d-94fb-4acc-b9ba-aa13e51191ff", 00:23:08.017 "is_configured": true, 00:23:08.017 "data_offset": 2048, 00:23:08.017 "data_size": 63488 00:23:08.017 }, 00:23:08.017 { 00:23:08.017 "name": null, 00:23:08.017 "uuid": "c93c52bb-c0ce-455c-9bca-bce75d2807ff", 00:23:08.017 "is_configured": false, 00:23:08.017 "data_offset": 0, 00:23:08.017 "data_size": 63488 00:23:08.017 }, 00:23:08.017 { 00:23:08.017 "name": null, 00:23:08.017 "uuid": "df971be6-3f78-4484-b8e6-d8807b482b40", 00:23:08.017 "is_configured": false, 00:23:08.017 "data_offset": 0, 00:23:08.017 "data_size": 63488 00:23:08.017 } 00:23:08.017 ] 00:23:08.017 }' 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:08.017 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:08.583 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:23:08.583 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:08.583 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:08.583 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:08.583 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:08.583 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:23:08.583 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:08.584 [2024-10-01 20:22:03.584258] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:08.584 "name": "Existed_Raid", 00:23:08.584 "uuid": "841c252b-cbdb-44df-b0d8-82ac6b43feb0", 00:23:08.584 "strip_size_kb": 0, 00:23:08.584 "state": "configuring", 00:23:08.584 "raid_level": "raid1", 00:23:08.584 "superblock": true, 00:23:08.584 "num_base_bdevs": 3, 00:23:08.584 "num_base_bdevs_discovered": 2, 00:23:08.584 "num_base_bdevs_operational": 3, 00:23:08.584 "base_bdevs_list": [ 00:23:08.584 { 00:23:08.584 "name": "BaseBdev1", 00:23:08.584 "uuid": "2b9ea74d-94fb-4acc-b9ba-aa13e51191ff", 00:23:08.584 "is_configured": true, 00:23:08.584 "data_offset": 2048, 00:23:08.584 "data_size": 63488 00:23:08.584 }, 00:23:08.584 { 00:23:08.584 "name": null, 00:23:08.584 "uuid": "c93c52bb-c0ce-455c-9bca-bce75d2807ff", 00:23:08.584 "is_configured": false, 00:23:08.584 "data_offset": 0, 00:23:08.584 "data_size": 63488 00:23:08.584 }, 00:23:08.584 { 00:23:08.584 "name": "BaseBdev3", 00:23:08.584 "uuid": "df971be6-3f78-4484-b8e6-d8807b482b40", 00:23:08.584 "is_configured": true, 00:23:08.584 "data_offset": 2048, 00:23:08.584 "data_size": 63488 00:23:08.584 } 00:23:08.584 ] 00:23:08.584 }' 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:08.584 20:22:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:08.842 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:08.842 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:23:08.842 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:08.842 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:09.101 [2024-10-01 20:22:04.132500] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:09.101 "name": "Existed_Raid", 00:23:09.101 "uuid": "841c252b-cbdb-44df-b0d8-82ac6b43feb0", 00:23:09.101 "strip_size_kb": 0, 00:23:09.101 "state": "configuring", 00:23:09.101 "raid_level": "raid1", 00:23:09.101 "superblock": true, 00:23:09.101 "num_base_bdevs": 3, 00:23:09.101 "num_base_bdevs_discovered": 1, 00:23:09.101 "num_base_bdevs_operational": 3, 00:23:09.101 "base_bdevs_list": [ 00:23:09.101 { 00:23:09.101 "name": null, 00:23:09.101 "uuid": "2b9ea74d-94fb-4acc-b9ba-aa13e51191ff", 00:23:09.101 "is_configured": false, 00:23:09.101 "data_offset": 0, 00:23:09.101 "data_size": 63488 00:23:09.101 }, 00:23:09.101 { 00:23:09.101 "name": null, 00:23:09.101 "uuid": "c93c52bb-c0ce-455c-9bca-bce75d2807ff", 00:23:09.101 "is_configured": false, 00:23:09.101 "data_offset": 0, 00:23:09.101 "data_size": 63488 00:23:09.101 }, 00:23:09.101 { 00:23:09.101 "name": "BaseBdev3", 00:23:09.101 "uuid": "df971be6-3f78-4484-b8e6-d8807b482b40", 00:23:09.101 "is_configured": true, 00:23:09.101 "data_offset": 2048, 00:23:09.101 "data_size": 63488 00:23:09.101 } 00:23:09.101 ] 00:23:09.101 }' 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:09.101 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:09.668 [2024-10-01 20:22:04.797692] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:09.668 "name": "Existed_Raid", 00:23:09.668 "uuid": "841c252b-cbdb-44df-b0d8-82ac6b43feb0", 00:23:09.668 "strip_size_kb": 0, 00:23:09.668 "state": "configuring", 00:23:09.668 "raid_level": "raid1", 00:23:09.668 "superblock": true, 00:23:09.668 "num_base_bdevs": 3, 00:23:09.668 "num_base_bdevs_discovered": 2, 00:23:09.668 "num_base_bdevs_operational": 3, 00:23:09.668 "base_bdevs_list": [ 00:23:09.668 { 00:23:09.668 "name": null, 00:23:09.668 "uuid": "2b9ea74d-94fb-4acc-b9ba-aa13e51191ff", 00:23:09.668 "is_configured": false, 00:23:09.668 "data_offset": 0, 00:23:09.668 "data_size": 63488 00:23:09.668 }, 00:23:09.668 { 00:23:09.668 "name": "BaseBdev2", 00:23:09.668 "uuid": "c93c52bb-c0ce-455c-9bca-bce75d2807ff", 00:23:09.668 "is_configured": true, 00:23:09.668 "data_offset": 2048, 00:23:09.668 "data_size": 63488 00:23:09.668 }, 00:23:09.668 { 00:23:09.668 "name": "BaseBdev3", 00:23:09.668 "uuid": "df971be6-3f78-4484-b8e6-d8807b482b40", 00:23:09.668 "is_configured": true, 00:23:09.668 "data_offset": 2048, 00:23:09.668 "data_size": 63488 00:23:09.668 } 00:23:09.668 ] 00:23:09.668 }' 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:09.668 20:22:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 2b9ea74d-94fb-4acc-b9ba-aa13e51191ff 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:10.235 [2024-10-01 20:22:05.467309] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:23:10.235 [2024-10-01 20:22:05.467771] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:23:10.235 [2024-10-01 20:22:05.467817] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:23:10.235 NewBaseBdev 00:23:10.235 [2024-10-01 20:22:05.468226] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:23:10.235 [2024-10-01 20:22:05.468487] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:23:10.235 [2024-10-01 20:22:05.468522] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:23:10.235 [2024-10-01 20:22:05.468716] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:10.235 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:10.236 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:10.236 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:10.236 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:10.236 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:10.236 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:23:10.236 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:10.236 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:10.236 [ 00:23:10.236 { 00:23:10.495 "name": "NewBaseBdev", 00:23:10.495 "aliases": [ 00:23:10.495 "2b9ea74d-94fb-4acc-b9ba-aa13e51191ff" 00:23:10.495 ], 00:23:10.495 "product_name": "Malloc disk", 00:23:10.495 "block_size": 512, 00:23:10.495 "num_blocks": 65536, 00:23:10.495 "uuid": "2b9ea74d-94fb-4acc-b9ba-aa13e51191ff", 00:23:10.495 "assigned_rate_limits": { 00:23:10.495 "rw_ios_per_sec": 0, 00:23:10.495 "rw_mbytes_per_sec": 0, 00:23:10.495 "r_mbytes_per_sec": 0, 00:23:10.495 "w_mbytes_per_sec": 0 00:23:10.495 }, 00:23:10.495 "claimed": true, 00:23:10.495 "claim_type": "exclusive_write", 00:23:10.495 "zoned": false, 00:23:10.495 "supported_io_types": { 00:23:10.495 "read": true, 00:23:10.496 "write": true, 00:23:10.496 "unmap": true, 00:23:10.496 "flush": true, 00:23:10.496 "reset": true, 00:23:10.496 "nvme_admin": false, 00:23:10.496 "nvme_io": false, 00:23:10.496 "nvme_io_md": false, 00:23:10.496 "write_zeroes": true, 00:23:10.496 "zcopy": true, 00:23:10.496 "get_zone_info": false, 00:23:10.496 "zone_management": false, 00:23:10.496 "zone_append": false, 00:23:10.496 "compare": false, 00:23:10.496 "compare_and_write": false, 00:23:10.496 "abort": true, 00:23:10.496 "seek_hole": false, 00:23:10.496 "seek_data": false, 00:23:10.496 "copy": true, 00:23:10.496 "nvme_iov_md": false 00:23:10.496 }, 00:23:10.496 "memory_domains": [ 00:23:10.496 { 00:23:10.496 "dma_device_id": "system", 00:23:10.496 "dma_device_type": 1 00:23:10.496 }, 00:23:10.496 { 00:23:10.496 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:10.496 "dma_device_type": 2 00:23:10.496 } 00:23:10.496 ], 00:23:10.496 "driver_specific": {} 00:23:10.496 } 00:23:10.496 ] 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:10.496 "name": "Existed_Raid", 00:23:10.496 "uuid": "841c252b-cbdb-44df-b0d8-82ac6b43feb0", 00:23:10.496 "strip_size_kb": 0, 00:23:10.496 "state": "online", 00:23:10.496 "raid_level": "raid1", 00:23:10.496 "superblock": true, 00:23:10.496 "num_base_bdevs": 3, 00:23:10.496 "num_base_bdevs_discovered": 3, 00:23:10.496 "num_base_bdevs_operational": 3, 00:23:10.496 "base_bdevs_list": [ 00:23:10.496 { 00:23:10.496 "name": "NewBaseBdev", 00:23:10.496 "uuid": "2b9ea74d-94fb-4acc-b9ba-aa13e51191ff", 00:23:10.496 "is_configured": true, 00:23:10.496 "data_offset": 2048, 00:23:10.496 "data_size": 63488 00:23:10.496 }, 00:23:10.496 { 00:23:10.496 "name": "BaseBdev2", 00:23:10.496 "uuid": "c93c52bb-c0ce-455c-9bca-bce75d2807ff", 00:23:10.496 "is_configured": true, 00:23:10.496 "data_offset": 2048, 00:23:10.496 "data_size": 63488 00:23:10.496 }, 00:23:10.496 { 00:23:10.496 "name": "BaseBdev3", 00:23:10.496 "uuid": "df971be6-3f78-4484-b8e6-d8807b482b40", 00:23:10.496 "is_configured": true, 00:23:10.496 "data_offset": 2048, 00:23:10.496 "data_size": 63488 00:23:10.496 } 00:23:10.496 ] 00:23:10.496 }' 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:10.496 20:22:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:23:11.065 [2024-10-01 20:22:06.040027] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:11.065 "name": "Existed_Raid", 00:23:11.065 "aliases": [ 00:23:11.065 "841c252b-cbdb-44df-b0d8-82ac6b43feb0" 00:23:11.065 ], 00:23:11.065 "product_name": "Raid Volume", 00:23:11.065 "block_size": 512, 00:23:11.065 "num_blocks": 63488, 00:23:11.065 "uuid": "841c252b-cbdb-44df-b0d8-82ac6b43feb0", 00:23:11.065 "assigned_rate_limits": { 00:23:11.065 "rw_ios_per_sec": 0, 00:23:11.065 "rw_mbytes_per_sec": 0, 00:23:11.065 "r_mbytes_per_sec": 0, 00:23:11.065 "w_mbytes_per_sec": 0 00:23:11.065 }, 00:23:11.065 "claimed": false, 00:23:11.065 "zoned": false, 00:23:11.065 "supported_io_types": { 00:23:11.065 "read": true, 00:23:11.065 "write": true, 00:23:11.065 "unmap": false, 00:23:11.065 "flush": false, 00:23:11.065 "reset": true, 00:23:11.065 "nvme_admin": false, 00:23:11.065 "nvme_io": false, 00:23:11.065 "nvme_io_md": false, 00:23:11.065 "write_zeroes": true, 00:23:11.065 "zcopy": false, 00:23:11.065 "get_zone_info": false, 00:23:11.065 "zone_management": false, 00:23:11.065 "zone_append": false, 00:23:11.065 "compare": false, 00:23:11.065 "compare_and_write": false, 00:23:11.065 "abort": false, 00:23:11.065 "seek_hole": false, 00:23:11.065 "seek_data": false, 00:23:11.065 "copy": false, 00:23:11.065 "nvme_iov_md": false 00:23:11.065 }, 00:23:11.065 "memory_domains": [ 00:23:11.065 { 00:23:11.065 "dma_device_id": "system", 00:23:11.065 "dma_device_type": 1 00:23:11.065 }, 00:23:11.065 { 00:23:11.065 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:11.065 "dma_device_type": 2 00:23:11.065 }, 00:23:11.065 { 00:23:11.065 "dma_device_id": "system", 00:23:11.065 "dma_device_type": 1 00:23:11.065 }, 00:23:11.065 { 00:23:11.065 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:11.065 "dma_device_type": 2 00:23:11.065 }, 00:23:11.065 { 00:23:11.065 "dma_device_id": "system", 00:23:11.065 "dma_device_type": 1 00:23:11.065 }, 00:23:11.065 { 00:23:11.065 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:11.065 "dma_device_type": 2 00:23:11.065 } 00:23:11.065 ], 00:23:11.065 "driver_specific": { 00:23:11.065 "raid": { 00:23:11.065 "uuid": "841c252b-cbdb-44df-b0d8-82ac6b43feb0", 00:23:11.065 "strip_size_kb": 0, 00:23:11.065 "state": "online", 00:23:11.065 "raid_level": "raid1", 00:23:11.065 "superblock": true, 00:23:11.065 "num_base_bdevs": 3, 00:23:11.065 "num_base_bdevs_discovered": 3, 00:23:11.065 "num_base_bdevs_operational": 3, 00:23:11.065 "base_bdevs_list": [ 00:23:11.065 { 00:23:11.065 "name": "NewBaseBdev", 00:23:11.065 "uuid": "2b9ea74d-94fb-4acc-b9ba-aa13e51191ff", 00:23:11.065 "is_configured": true, 00:23:11.065 "data_offset": 2048, 00:23:11.065 "data_size": 63488 00:23:11.065 }, 00:23:11.065 { 00:23:11.065 "name": "BaseBdev2", 00:23:11.065 "uuid": "c93c52bb-c0ce-455c-9bca-bce75d2807ff", 00:23:11.065 "is_configured": true, 00:23:11.065 "data_offset": 2048, 00:23:11.065 "data_size": 63488 00:23:11.065 }, 00:23:11.065 { 00:23:11.065 "name": "BaseBdev3", 00:23:11.065 "uuid": "df971be6-3f78-4484-b8e6-d8807b482b40", 00:23:11.065 "is_configured": true, 00:23:11.065 "data_offset": 2048, 00:23:11.065 "data_size": 63488 00:23:11.065 } 00:23:11.065 ] 00:23:11.065 } 00:23:11.065 } 00:23:11.065 }' 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:23:11.065 BaseBdev2 00:23:11.065 BaseBdev3' 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:11.065 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:11.324 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:11.324 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:11.324 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:23:11.324 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:11.324 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:11.324 [2024-10-01 20:22:06.355581] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:23:11.324 [2024-10-01 20:22:06.355643] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:23:11.324 [2024-10-01 20:22:06.355777] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:23:11.324 [2024-10-01 20:22:06.356207] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:23:11.324 [2024-10-01 20:22:06.356250] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:23:11.324 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:11.324 20:22:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 68683 00:23:11.324 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 68683 ']' 00:23:11.324 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 68683 00:23:11.324 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:23:11.325 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:11.325 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 68683 00:23:11.325 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:23:11.325 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:23:11.325 killing process with pid 68683 00:23:11.325 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 68683' 00:23:11.325 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 68683 00:23:11.325 20:22:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 68683 00:23:11.325 [2024-10-01 20:22:06.395530] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:23:11.583 [2024-10-01 20:22:06.661059] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:23:13.485 20:22:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:23:13.485 00:23:13.485 real 0m12.745s 00:23:13.485 user 0m20.360s 00:23:13.485 sys 0m1.778s 00:23:13.485 20:22:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:23:13.485 20:22:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:13.485 ************************************ 00:23:13.485 END TEST raid_state_function_test_sb 00:23:13.485 ************************************ 00:23:13.485 20:22:08 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 3 00:23:13.485 20:22:08 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:23:13.485 20:22:08 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:23:13.485 20:22:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:23:13.485 ************************************ 00:23:13.485 START TEST raid_superblock_test 00:23:13.485 ************************************ 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 3 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=69326 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 69326 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 69326 ']' 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:13.485 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:13.485 20:22:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:13.485 [2024-10-01 20:22:08.623790] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:23:13.485 [2024-10-01 20:22:08.623961] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69326 ] 00:23:13.744 [2024-10-01 20:22:08.791453] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:14.001 [2024-10-01 20:22:09.038224] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:23:14.001 [2024-10-01 20:22:09.244562] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:23:14.001 [2024-10-01 20:22:09.244650] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:14.567 malloc1 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:14.567 [2024-10-01 20:22:09.709679] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:23:14.567 [2024-10-01 20:22:09.709772] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:14.567 [2024-10-01 20:22:09.709810] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:23:14.567 [2024-10-01 20:22:09.709830] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:14.567 [2024-10-01 20:22:09.712748] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:14.567 [2024-10-01 20:22:09.712793] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:23:14.567 pt1 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:14.567 malloc2 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:14.567 [2024-10-01 20:22:09.758635] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:23:14.567 [2024-10-01 20:22:09.758737] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:14.567 [2024-10-01 20:22:09.758779] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:23:14.567 [2024-10-01 20:22:09.758796] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:14.567 [2024-10-01 20:22:09.762112] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:14.567 [2024-10-01 20:22:09.762202] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:23:14.567 pt2 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:14.567 malloc3 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:14.567 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:14.567 [2024-10-01 20:22:09.812261] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:23:14.567 [2024-10-01 20:22:09.812329] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:14.567 [2024-10-01 20:22:09.812361] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:23:14.567 [2024-10-01 20:22:09.812378] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:14.568 [2024-10-01 20:22:09.815270] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:14.568 [2024-10-01 20:22:09.815330] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:23:14.568 pt3 00:23:14.568 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:14.568 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:23:14.568 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:23:14.568 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:23:14.568 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:14.568 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:14.826 [2024-10-01 20:22:09.820335] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:23:14.826 [2024-10-01 20:22:09.822971] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:23:14.826 [2024-10-01 20:22:09.823085] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:23:14.826 [2024-10-01 20:22:09.823326] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:23:14.826 [2024-10-01 20:22:09.823360] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:23:14.826 [2024-10-01 20:22:09.823667] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:23:14.826 [2024-10-01 20:22:09.823928] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:23:14.826 [2024-10-01 20:22:09.823955] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:23:14.826 [2024-10-01 20:22:09.824138] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:14.826 "name": "raid_bdev1", 00:23:14.826 "uuid": "e4b926a5-0f83-42e3-873f-73d49577e743", 00:23:14.826 "strip_size_kb": 0, 00:23:14.826 "state": "online", 00:23:14.826 "raid_level": "raid1", 00:23:14.826 "superblock": true, 00:23:14.826 "num_base_bdevs": 3, 00:23:14.826 "num_base_bdevs_discovered": 3, 00:23:14.826 "num_base_bdevs_operational": 3, 00:23:14.826 "base_bdevs_list": [ 00:23:14.826 { 00:23:14.826 "name": "pt1", 00:23:14.826 "uuid": "00000000-0000-0000-0000-000000000001", 00:23:14.826 "is_configured": true, 00:23:14.826 "data_offset": 2048, 00:23:14.826 "data_size": 63488 00:23:14.826 }, 00:23:14.826 { 00:23:14.826 "name": "pt2", 00:23:14.826 "uuid": "00000000-0000-0000-0000-000000000002", 00:23:14.826 "is_configured": true, 00:23:14.826 "data_offset": 2048, 00:23:14.826 "data_size": 63488 00:23:14.826 }, 00:23:14.826 { 00:23:14.826 "name": "pt3", 00:23:14.826 "uuid": "00000000-0000-0000-0000-000000000003", 00:23:14.826 "is_configured": true, 00:23:14.826 "data_offset": 2048, 00:23:14.826 "data_size": 63488 00:23:14.826 } 00:23:14.826 ] 00:23:14.826 }' 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:14.826 20:22:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:15.085 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:23:15.085 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:23:15.085 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:23:15.085 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:23:15.085 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:23:15.085 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:23:15.085 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:23:15.085 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:23:15.085 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:15.085 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:15.085 [2024-10-01 20:22:10.321000] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:23:15.343 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:15.343 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:15.343 "name": "raid_bdev1", 00:23:15.343 "aliases": [ 00:23:15.343 "e4b926a5-0f83-42e3-873f-73d49577e743" 00:23:15.343 ], 00:23:15.343 "product_name": "Raid Volume", 00:23:15.343 "block_size": 512, 00:23:15.343 "num_blocks": 63488, 00:23:15.343 "uuid": "e4b926a5-0f83-42e3-873f-73d49577e743", 00:23:15.343 "assigned_rate_limits": { 00:23:15.343 "rw_ios_per_sec": 0, 00:23:15.343 "rw_mbytes_per_sec": 0, 00:23:15.343 "r_mbytes_per_sec": 0, 00:23:15.343 "w_mbytes_per_sec": 0 00:23:15.343 }, 00:23:15.343 "claimed": false, 00:23:15.343 "zoned": false, 00:23:15.343 "supported_io_types": { 00:23:15.343 "read": true, 00:23:15.343 "write": true, 00:23:15.343 "unmap": false, 00:23:15.343 "flush": false, 00:23:15.343 "reset": true, 00:23:15.343 "nvme_admin": false, 00:23:15.343 "nvme_io": false, 00:23:15.343 "nvme_io_md": false, 00:23:15.343 "write_zeroes": true, 00:23:15.343 "zcopy": false, 00:23:15.343 "get_zone_info": false, 00:23:15.343 "zone_management": false, 00:23:15.343 "zone_append": false, 00:23:15.343 "compare": false, 00:23:15.343 "compare_and_write": false, 00:23:15.343 "abort": false, 00:23:15.343 "seek_hole": false, 00:23:15.343 "seek_data": false, 00:23:15.343 "copy": false, 00:23:15.343 "nvme_iov_md": false 00:23:15.343 }, 00:23:15.343 "memory_domains": [ 00:23:15.343 { 00:23:15.343 "dma_device_id": "system", 00:23:15.343 "dma_device_type": 1 00:23:15.343 }, 00:23:15.343 { 00:23:15.343 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:15.343 "dma_device_type": 2 00:23:15.343 }, 00:23:15.343 { 00:23:15.343 "dma_device_id": "system", 00:23:15.343 "dma_device_type": 1 00:23:15.343 }, 00:23:15.343 { 00:23:15.343 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:15.343 "dma_device_type": 2 00:23:15.343 }, 00:23:15.343 { 00:23:15.343 "dma_device_id": "system", 00:23:15.343 "dma_device_type": 1 00:23:15.343 }, 00:23:15.343 { 00:23:15.343 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:15.343 "dma_device_type": 2 00:23:15.343 } 00:23:15.343 ], 00:23:15.343 "driver_specific": { 00:23:15.343 "raid": { 00:23:15.343 "uuid": "e4b926a5-0f83-42e3-873f-73d49577e743", 00:23:15.343 "strip_size_kb": 0, 00:23:15.343 "state": "online", 00:23:15.343 "raid_level": "raid1", 00:23:15.343 "superblock": true, 00:23:15.343 "num_base_bdevs": 3, 00:23:15.343 "num_base_bdevs_discovered": 3, 00:23:15.343 "num_base_bdevs_operational": 3, 00:23:15.343 "base_bdevs_list": [ 00:23:15.343 { 00:23:15.343 "name": "pt1", 00:23:15.343 "uuid": "00000000-0000-0000-0000-000000000001", 00:23:15.343 "is_configured": true, 00:23:15.343 "data_offset": 2048, 00:23:15.343 "data_size": 63488 00:23:15.343 }, 00:23:15.343 { 00:23:15.343 "name": "pt2", 00:23:15.343 "uuid": "00000000-0000-0000-0000-000000000002", 00:23:15.343 "is_configured": true, 00:23:15.343 "data_offset": 2048, 00:23:15.343 "data_size": 63488 00:23:15.343 }, 00:23:15.343 { 00:23:15.343 "name": "pt3", 00:23:15.343 "uuid": "00000000-0000-0000-0000-000000000003", 00:23:15.343 "is_configured": true, 00:23:15.343 "data_offset": 2048, 00:23:15.343 "data_size": 63488 00:23:15.343 } 00:23:15.343 ] 00:23:15.343 } 00:23:15.343 } 00:23:15.343 }' 00:23:15.343 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:23:15.343 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:23:15.343 pt2 00:23:15.343 pt3' 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:15.344 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:23:15.602 [2024-10-01 20:22:10.621044] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=e4b926a5-0f83-42e3-873f-73d49577e743 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z e4b926a5-0f83-42e3-873f-73d49577e743 ']' 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:15.602 [2024-10-01 20:22:10.668627] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:23:15.602 [2024-10-01 20:22:10.668664] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:23:15.602 [2024-10-01 20:22:10.668795] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:23:15.602 [2024-10-01 20:22:10.668929] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:23:15.602 [2024-10-01 20:22:10.668948] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:15.602 [2024-10-01 20:22:10.824753] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:23:15.602 [2024-10-01 20:22:10.827488] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:23:15.602 [2024-10-01 20:22:10.827562] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:23:15.602 [2024-10-01 20:22:10.827654] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:23:15.602 [2024-10-01 20:22:10.827757] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:23:15.602 [2024-10-01 20:22:10.827796] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:23:15.602 [2024-10-01 20:22:10.827826] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:23:15.602 [2024-10-01 20:22:10.827840] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:23:15.602 request: 00:23:15.602 { 00:23:15.602 "name": "raid_bdev1", 00:23:15.602 "raid_level": "raid1", 00:23:15.602 "base_bdevs": [ 00:23:15.602 "malloc1", 00:23:15.602 "malloc2", 00:23:15.602 "malloc3" 00:23:15.602 ], 00:23:15.602 "superblock": false, 00:23:15.602 "method": "bdev_raid_create", 00:23:15.602 "req_id": 1 00:23:15.602 } 00:23:15.602 Got JSON-RPC error response 00:23:15.602 response: 00:23:15.602 { 00:23:15.602 "code": -17, 00:23:15.602 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:23:15.602 } 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:23:15.602 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:15.603 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:15.603 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:23:15.603 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:15.603 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:15.860 [2024-10-01 20:22:10.892732] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:23:15.860 [2024-10-01 20:22:10.892963] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:15.860 [2024-10-01 20:22:10.893043] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:23:15.860 [2024-10-01 20:22:10.893255] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:15.860 [2024-10-01 20:22:10.896368] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:15.860 [2024-10-01 20:22:10.896530] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:23:15.860 [2024-10-01 20:22:10.896790] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:23:15.860 [2024-10-01 20:22:10.896984] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:23:15.860 pt1 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:15.860 "name": "raid_bdev1", 00:23:15.860 "uuid": "e4b926a5-0f83-42e3-873f-73d49577e743", 00:23:15.860 "strip_size_kb": 0, 00:23:15.860 "state": "configuring", 00:23:15.860 "raid_level": "raid1", 00:23:15.860 "superblock": true, 00:23:15.860 "num_base_bdevs": 3, 00:23:15.860 "num_base_bdevs_discovered": 1, 00:23:15.860 "num_base_bdevs_operational": 3, 00:23:15.860 "base_bdevs_list": [ 00:23:15.860 { 00:23:15.860 "name": "pt1", 00:23:15.860 "uuid": "00000000-0000-0000-0000-000000000001", 00:23:15.860 "is_configured": true, 00:23:15.860 "data_offset": 2048, 00:23:15.860 "data_size": 63488 00:23:15.860 }, 00:23:15.860 { 00:23:15.860 "name": null, 00:23:15.860 "uuid": "00000000-0000-0000-0000-000000000002", 00:23:15.860 "is_configured": false, 00:23:15.860 "data_offset": 2048, 00:23:15.860 "data_size": 63488 00:23:15.860 }, 00:23:15.860 { 00:23:15.860 "name": null, 00:23:15.860 "uuid": "00000000-0000-0000-0000-000000000003", 00:23:15.860 "is_configured": false, 00:23:15.860 "data_offset": 2048, 00:23:15.860 "data_size": 63488 00:23:15.860 } 00:23:15.860 ] 00:23:15.860 }' 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:15.860 20:22:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:16.427 [2024-10-01 20:22:11.401093] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:23:16.427 [2024-10-01 20:22:11.401325] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:16.427 [2024-10-01 20:22:11.401375] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:23:16.427 [2024-10-01 20:22:11.401394] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:16.427 [2024-10-01 20:22:11.402003] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:16.427 [2024-10-01 20:22:11.402036] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:23:16.427 [2024-10-01 20:22:11.402155] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:23:16.427 [2024-10-01 20:22:11.402188] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:23:16.427 pt2 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:16.427 [2024-10-01 20:22:11.409061] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:16.427 "name": "raid_bdev1", 00:23:16.427 "uuid": "e4b926a5-0f83-42e3-873f-73d49577e743", 00:23:16.427 "strip_size_kb": 0, 00:23:16.427 "state": "configuring", 00:23:16.427 "raid_level": "raid1", 00:23:16.427 "superblock": true, 00:23:16.427 "num_base_bdevs": 3, 00:23:16.427 "num_base_bdevs_discovered": 1, 00:23:16.427 "num_base_bdevs_operational": 3, 00:23:16.427 "base_bdevs_list": [ 00:23:16.427 { 00:23:16.427 "name": "pt1", 00:23:16.427 "uuid": "00000000-0000-0000-0000-000000000001", 00:23:16.427 "is_configured": true, 00:23:16.427 "data_offset": 2048, 00:23:16.427 "data_size": 63488 00:23:16.427 }, 00:23:16.427 { 00:23:16.427 "name": null, 00:23:16.427 "uuid": "00000000-0000-0000-0000-000000000002", 00:23:16.427 "is_configured": false, 00:23:16.427 "data_offset": 0, 00:23:16.427 "data_size": 63488 00:23:16.427 }, 00:23:16.427 { 00:23:16.427 "name": null, 00:23:16.427 "uuid": "00000000-0000-0000-0000-000000000003", 00:23:16.427 "is_configured": false, 00:23:16.427 "data_offset": 2048, 00:23:16.427 "data_size": 63488 00:23:16.427 } 00:23:16.427 ] 00:23:16.427 }' 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:16.427 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:16.993 [2024-10-01 20:22:11.969222] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:23:16.993 [2024-10-01 20:22:11.969332] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:16.993 [2024-10-01 20:22:11.969373] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:23:16.993 [2024-10-01 20:22:11.969392] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:16.993 [2024-10-01 20:22:11.970033] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:16.993 [2024-10-01 20:22:11.970065] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:23:16.993 [2024-10-01 20:22:11.970193] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:23:16.993 [2024-10-01 20:22:11.970241] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:23:16.993 pt2 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:16.993 [2024-10-01 20:22:11.977189] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:23:16.993 [2024-10-01 20:22:11.977250] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:16.993 [2024-10-01 20:22:11.977273] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:23:16.993 [2024-10-01 20:22:11.977289] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:16.993 [2024-10-01 20:22:11.977805] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:16.993 [2024-10-01 20:22:11.977846] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:23:16.993 [2024-10-01 20:22:11.977929] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:23:16.993 [2024-10-01 20:22:11.977969] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:23:16.993 [2024-10-01 20:22:11.978119] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:23:16.993 [2024-10-01 20:22:11.978143] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:23:16.993 [2024-10-01 20:22:11.978452] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:23:16.993 [2024-10-01 20:22:11.978656] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:23:16.993 [2024-10-01 20:22:11.978673] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:23:16.993 [2024-10-01 20:22:11.978861] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:16.993 pt3 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:16.993 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:16.994 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:16.994 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:16.994 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:16.994 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:16.994 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:16.994 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:16.994 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:16.994 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:16.994 20:22:11 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:16.994 20:22:11 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:16.994 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:16.994 "name": "raid_bdev1", 00:23:16.994 "uuid": "e4b926a5-0f83-42e3-873f-73d49577e743", 00:23:16.994 "strip_size_kb": 0, 00:23:16.994 "state": "online", 00:23:16.994 "raid_level": "raid1", 00:23:16.994 "superblock": true, 00:23:16.994 "num_base_bdevs": 3, 00:23:16.994 "num_base_bdevs_discovered": 3, 00:23:16.994 "num_base_bdevs_operational": 3, 00:23:16.994 "base_bdevs_list": [ 00:23:16.994 { 00:23:16.994 "name": "pt1", 00:23:16.994 "uuid": "00000000-0000-0000-0000-000000000001", 00:23:16.994 "is_configured": true, 00:23:16.994 "data_offset": 2048, 00:23:16.994 "data_size": 63488 00:23:16.994 }, 00:23:16.994 { 00:23:16.994 "name": "pt2", 00:23:16.994 "uuid": "00000000-0000-0000-0000-000000000002", 00:23:16.994 "is_configured": true, 00:23:16.994 "data_offset": 2048, 00:23:16.994 "data_size": 63488 00:23:16.994 }, 00:23:16.994 { 00:23:16.994 "name": "pt3", 00:23:16.994 "uuid": "00000000-0000-0000-0000-000000000003", 00:23:16.994 "is_configured": true, 00:23:16.994 "data_offset": 2048, 00:23:16.994 "data_size": 63488 00:23:16.994 } 00:23:16.994 ] 00:23:16.994 }' 00:23:16.994 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:16.994 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:17.253 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:23:17.253 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:23:17.253 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:23:17.253 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:23:17.253 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:23:17.253 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:23:17.253 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:23:17.253 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:23:17.253 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:17.253 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:17.253 [2024-10-01 20:22:12.501761] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:23:17.511 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:17.511 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:17.511 "name": "raid_bdev1", 00:23:17.512 "aliases": [ 00:23:17.512 "e4b926a5-0f83-42e3-873f-73d49577e743" 00:23:17.512 ], 00:23:17.512 "product_name": "Raid Volume", 00:23:17.512 "block_size": 512, 00:23:17.512 "num_blocks": 63488, 00:23:17.512 "uuid": "e4b926a5-0f83-42e3-873f-73d49577e743", 00:23:17.512 "assigned_rate_limits": { 00:23:17.512 "rw_ios_per_sec": 0, 00:23:17.512 "rw_mbytes_per_sec": 0, 00:23:17.512 "r_mbytes_per_sec": 0, 00:23:17.512 "w_mbytes_per_sec": 0 00:23:17.512 }, 00:23:17.512 "claimed": false, 00:23:17.512 "zoned": false, 00:23:17.512 "supported_io_types": { 00:23:17.512 "read": true, 00:23:17.512 "write": true, 00:23:17.512 "unmap": false, 00:23:17.512 "flush": false, 00:23:17.512 "reset": true, 00:23:17.512 "nvme_admin": false, 00:23:17.512 "nvme_io": false, 00:23:17.512 "nvme_io_md": false, 00:23:17.512 "write_zeroes": true, 00:23:17.512 "zcopy": false, 00:23:17.512 "get_zone_info": false, 00:23:17.512 "zone_management": false, 00:23:17.512 "zone_append": false, 00:23:17.512 "compare": false, 00:23:17.512 "compare_and_write": false, 00:23:17.512 "abort": false, 00:23:17.512 "seek_hole": false, 00:23:17.512 "seek_data": false, 00:23:17.512 "copy": false, 00:23:17.512 "nvme_iov_md": false 00:23:17.512 }, 00:23:17.512 "memory_domains": [ 00:23:17.512 { 00:23:17.512 "dma_device_id": "system", 00:23:17.512 "dma_device_type": 1 00:23:17.512 }, 00:23:17.512 { 00:23:17.512 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:17.512 "dma_device_type": 2 00:23:17.512 }, 00:23:17.512 { 00:23:17.512 "dma_device_id": "system", 00:23:17.512 "dma_device_type": 1 00:23:17.512 }, 00:23:17.512 { 00:23:17.512 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:17.512 "dma_device_type": 2 00:23:17.512 }, 00:23:17.512 { 00:23:17.512 "dma_device_id": "system", 00:23:17.512 "dma_device_type": 1 00:23:17.512 }, 00:23:17.512 { 00:23:17.512 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:17.512 "dma_device_type": 2 00:23:17.512 } 00:23:17.512 ], 00:23:17.512 "driver_specific": { 00:23:17.512 "raid": { 00:23:17.512 "uuid": "e4b926a5-0f83-42e3-873f-73d49577e743", 00:23:17.512 "strip_size_kb": 0, 00:23:17.512 "state": "online", 00:23:17.512 "raid_level": "raid1", 00:23:17.512 "superblock": true, 00:23:17.512 "num_base_bdevs": 3, 00:23:17.512 "num_base_bdevs_discovered": 3, 00:23:17.512 "num_base_bdevs_operational": 3, 00:23:17.512 "base_bdevs_list": [ 00:23:17.512 { 00:23:17.512 "name": "pt1", 00:23:17.512 "uuid": "00000000-0000-0000-0000-000000000001", 00:23:17.512 "is_configured": true, 00:23:17.512 "data_offset": 2048, 00:23:17.512 "data_size": 63488 00:23:17.512 }, 00:23:17.512 { 00:23:17.512 "name": "pt2", 00:23:17.512 "uuid": "00000000-0000-0000-0000-000000000002", 00:23:17.512 "is_configured": true, 00:23:17.512 "data_offset": 2048, 00:23:17.512 "data_size": 63488 00:23:17.512 }, 00:23:17.512 { 00:23:17.512 "name": "pt3", 00:23:17.512 "uuid": "00000000-0000-0000-0000-000000000003", 00:23:17.512 "is_configured": true, 00:23:17.512 "data_offset": 2048, 00:23:17.512 "data_size": 63488 00:23:17.512 } 00:23:17.512 ] 00:23:17.512 } 00:23:17.512 } 00:23:17.512 }' 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:23:17.512 pt2 00:23:17.512 pt3' 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:17.512 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:17.771 [2024-10-01 20:22:12.821802] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' e4b926a5-0f83-42e3-873f-73d49577e743 '!=' e4b926a5-0f83-42e3-873f-73d49577e743 ']' 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:17.771 [2024-10-01 20:22:12.869543] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:17.771 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:17.771 "name": "raid_bdev1", 00:23:17.771 "uuid": "e4b926a5-0f83-42e3-873f-73d49577e743", 00:23:17.771 "strip_size_kb": 0, 00:23:17.771 "state": "online", 00:23:17.771 "raid_level": "raid1", 00:23:17.771 "superblock": true, 00:23:17.771 "num_base_bdevs": 3, 00:23:17.772 "num_base_bdevs_discovered": 2, 00:23:17.772 "num_base_bdevs_operational": 2, 00:23:17.772 "base_bdevs_list": [ 00:23:17.772 { 00:23:17.772 "name": null, 00:23:17.772 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:17.772 "is_configured": false, 00:23:17.772 "data_offset": 0, 00:23:17.772 "data_size": 63488 00:23:17.772 }, 00:23:17.772 { 00:23:17.772 "name": "pt2", 00:23:17.772 "uuid": "00000000-0000-0000-0000-000000000002", 00:23:17.772 "is_configured": true, 00:23:17.772 "data_offset": 2048, 00:23:17.772 "data_size": 63488 00:23:17.772 }, 00:23:17.772 { 00:23:17.772 "name": "pt3", 00:23:17.772 "uuid": "00000000-0000-0000-0000-000000000003", 00:23:17.772 "is_configured": true, 00:23:17.772 "data_offset": 2048, 00:23:17.772 "data_size": 63488 00:23:17.772 } 00:23:17.772 ] 00:23:17.772 }' 00:23:17.772 20:22:12 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:17.772 20:22:12 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:18.339 [2024-10-01 20:22:13.389663] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:23:18.339 [2024-10-01 20:22:13.389702] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:23:18.339 [2024-10-01 20:22:13.389820] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:23:18.339 [2024-10-01 20:22:13.389902] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:23:18.339 [2024-10-01 20:22:13.389926] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:18.339 [2024-10-01 20:22:13.469618] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:23:18.339 [2024-10-01 20:22:13.469697] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:18.339 [2024-10-01 20:22:13.469738] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:23:18.339 [2024-10-01 20:22:13.469760] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:18.339 [2024-10-01 20:22:13.472742] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:18.339 [2024-10-01 20:22:13.472787] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:23:18.339 [2024-10-01 20:22:13.472904] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:23:18.339 [2024-10-01 20:22:13.472976] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:23:18.339 pt2 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:18.339 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:18.339 "name": "raid_bdev1", 00:23:18.340 "uuid": "e4b926a5-0f83-42e3-873f-73d49577e743", 00:23:18.340 "strip_size_kb": 0, 00:23:18.340 "state": "configuring", 00:23:18.340 "raid_level": "raid1", 00:23:18.340 "superblock": true, 00:23:18.340 "num_base_bdevs": 3, 00:23:18.340 "num_base_bdevs_discovered": 1, 00:23:18.340 "num_base_bdevs_operational": 2, 00:23:18.340 "base_bdevs_list": [ 00:23:18.340 { 00:23:18.340 "name": null, 00:23:18.340 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:18.340 "is_configured": false, 00:23:18.340 "data_offset": 2048, 00:23:18.340 "data_size": 63488 00:23:18.340 }, 00:23:18.340 { 00:23:18.340 "name": "pt2", 00:23:18.340 "uuid": "00000000-0000-0000-0000-000000000002", 00:23:18.340 "is_configured": true, 00:23:18.340 "data_offset": 2048, 00:23:18.340 "data_size": 63488 00:23:18.340 }, 00:23:18.340 { 00:23:18.340 "name": null, 00:23:18.340 "uuid": "00000000-0000-0000-0000-000000000003", 00:23:18.340 "is_configured": false, 00:23:18.340 "data_offset": 2048, 00:23:18.340 "data_size": 63488 00:23:18.340 } 00:23:18.340 ] 00:23:18.340 }' 00:23:18.340 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:18.340 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:18.906 [2024-10-01 20:22:13.985874] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:23:18.906 [2024-10-01 20:22:13.986104] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:18.906 [2024-10-01 20:22:13.986148] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:23:18.906 [2024-10-01 20:22:13.986169] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:18.906 [2024-10-01 20:22:13.986825] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:18.906 [2024-10-01 20:22:13.986858] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:23:18.906 [2024-10-01 20:22:13.986971] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:23:18.906 [2024-10-01 20:22:13.987016] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:23:18.906 [2024-10-01 20:22:13.987163] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:23:18.906 [2024-10-01 20:22:13.987192] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:23:18.906 [2024-10-01 20:22:13.987493] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:23:18.906 [2024-10-01 20:22:13.987722] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:23:18.906 [2024-10-01 20:22:13.987771] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:23:18.906 [2024-10-01 20:22:13.987949] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:18.906 pt3 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:18.906 20:22:13 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:18.906 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:18.906 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:18.906 "name": "raid_bdev1", 00:23:18.906 "uuid": "e4b926a5-0f83-42e3-873f-73d49577e743", 00:23:18.906 "strip_size_kb": 0, 00:23:18.906 "state": "online", 00:23:18.906 "raid_level": "raid1", 00:23:18.906 "superblock": true, 00:23:18.906 "num_base_bdevs": 3, 00:23:18.906 "num_base_bdevs_discovered": 2, 00:23:18.906 "num_base_bdevs_operational": 2, 00:23:18.906 "base_bdevs_list": [ 00:23:18.906 { 00:23:18.906 "name": null, 00:23:18.906 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:18.906 "is_configured": false, 00:23:18.906 "data_offset": 2048, 00:23:18.906 "data_size": 63488 00:23:18.906 }, 00:23:18.906 { 00:23:18.906 "name": "pt2", 00:23:18.906 "uuid": "00000000-0000-0000-0000-000000000002", 00:23:18.906 "is_configured": true, 00:23:18.906 "data_offset": 2048, 00:23:18.906 "data_size": 63488 00:23:18.906 }, 00:23:18.906 { 00:23:18.906 "name": "pt3", 00:23:18.906 "uuid": "00000000-0000-0000-0000-000000000003", 00:23:18.906 "is_configured": true, 00:23:18.906 "data_offset": 2048, 00:23:18.906 "data_size": 63488 00:23:18.906 } 00:23:18.906 ] 00:23:18.906 }' 00:23:18.906 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:18.906 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:19.472 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:23:19.472 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:19.472 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:19.472 [2024-10-01 20:22:14.489965] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:23:19.472 [2024-10-01 20:22:14.490024] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:23:19.472 [2024-10-01 20:22:14.490187] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:23:19.472 [2024-10-01 20:22:14.490274] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:23:19.472 [2024-10-01 20:22:14.490290] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:23:19.472 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:19.472 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:23:19.472 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:19.472 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:19.472 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:19.472 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:19.472 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:23:19.472 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:19.473 [2024-10-01 20:22:14.562024] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:23:19.473 [2024-10-01 20:22:14.562249] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:19.473 [2024-10-01 20:22:14.562402] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:23:19.473 [2024-10-01 20:22:14.562427] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:19.473 [2024-10-01 20:22:14.565562] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:19.473 [2024-10-01 20:22:14.565607] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:23:19.473 [2024-10-01 20:22:14.565742] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:23:19.473 [2024-10-01 20:22:14.565804] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:23:19.473 [2024-10-01 20:22:14.566013] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:23:19.473 [2024-10-01 20:22:14.566036] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:23:19.473 [2024-10-01 20:22:14.566063] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:23:19.473 [2024-10-01 20:22:14.566129] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:23:19.473 pt1 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:19.473 "name": "raid_bdev1", 00:23:19.473 "uuid": "e4b926a5-0f83-42e3-873f-73d49577e743", 00:23:19.473 "strip_size_kb": 0, 00:23:19.473 "state": "configuring", 00:23:19.473 "raid_level": "raid1", 00:23:19.473 "superblock": true, 00:23:19.473 "num_base_bdevs": 3, 00:23:19.473 "num_base_bdevs_discovered": 1, 00:23:19.473 "num_base_bdevs_operational": 2, 00:23:19.473 "base_bdevs_list": [ 00:23:19.473 { 00:23:19.473 "name": null, 00:23:19.473 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:19.473 "is_configured": false, 00:23:19.473 "data_offset": 2048, 00:23:19.473 "data_size": 63488 00:23:19.473 }, 00:23:19.473 { 00:23:19.473 "name": "pt2", 00:23:19.473 "uuid": "00000000-0000-0000-0000-000000000002", 00:23:19.473 "is_configured": true, 00:23:19.473 "data_offset": 2048, 00:23:19.473 "data_size": 63488 00:23:19.473 }, 00:23:19.473 { 00:23:19.473 "name": null, 00:23:19.473 "uuid": "00000000-0000-0000-0000-000000000003", 00:23:19.473 "is_configured": false, 00:23:19.473 "data_offset": 2048, 00:23:19.473 "data_size": 63488 00:23:19.473 } 00:23:19.473 ] 00:23:19.473 }' 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:19.473 20:22:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:20.040 [2024-10-01 20:22:15.122304] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:23:20.040 [2024-10-01 20:22:15.122549] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:20.040 [2024-10-01 20:22:15.122598] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:23:20.040 [2024-10-01 20:22:15.122616] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:20.040 [2024-10-01 20:22:15.123227] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:20.040 [2024-10-01 20:22:15.123253] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:23:20.040 [2024-10-01 20:22:15.123369] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:23:20.040 [2024-10-01 20:22:15.123432] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:23:20.040 [2024-10-01 20:22:15.123595] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:23:20.040 [2024-10-01 20:22:15.123611] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:23:20.040 [2024-10-01 20:22:15.123989] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:23:20.040 [2024-10-01 20:22:15.124196] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:23:20.040 [2024-10-01 20:22:15.124226] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:23:20.040 [2024-10-01 20:22:15.124399] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:20.040 pt3 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:20.040 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:20.041 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:20.041 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:20.041 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:20.041 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:20.041 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:20.041 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:20.041 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:20.041 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:20.041 "name": "raid_bdev1", 00:23:20.041 "uuid": "e4b926a5-0f83-42e3-873f-73d49577e743", 00:23:20.041 "strip_size_kb": 0, 00:23:20.041 "state": "online", 00:23:20.041 "raid_level": "raid1", 00:23:20.041 "superblock": true, 00:23:20.041 "num_base_bdevs": 3, 00:23:20.041 "num_base_bdevs_discovered": 2, 00:23:20.041 "num_base_bdevs_operational": 2, 00:23:20.041 "base_bdevs_list": [ 00:23:20.041 { 00:23:20.041 "name": null, 00:23:20.041 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:20.041 "is_configured": false, 00:23:20.041 "data_offset": 2048, 00:23:20.041 "data_size": 63488 00:23:20.041 }, 00:23:20.041 { 00:23:20.041 "name": "pt2", 00:23:20.041 "uuid": "00000000-0000-0000-0000-000000000002", 00:23:20.041 "is_configured": true, 00:23:20.041 "data_offset": 2048, 00:23:20.041 "data_size": 63488 00:23:20.041 }, 00:23:20.041 { 00:23:20.041 "name": "pt3", 00:23:20.041 "uuid": "00000000-0000-0000-0000-000000000003", 00:23:20.041 "is_configured": true, 00:23:20.041 "data_offset": 2048, 00:23:20.041 "data_size": 63488 00:23:20.041 } 00:23:20.041 ] 00:23:20.041 }' 00:23:20.041 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:20.041 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:20.607 [2024-10-01 20:22:15.698789] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' e4b926a5-0f83-42e3-873f-73d49577e743 '!=' e4b926a5-0f83-42e3-873f-73d49577e743 ']' 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 69326 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 69326 ']' 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 69326 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69326 00:23:20.607 killing process with pid 69326 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69326' 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 69326 00:23:20.607 [2024-10-01 20:22:15.778147] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:23:20.607 20:22:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 69326 00:23:20.607 [2024-10-01 20:22:15.778271] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:23:20.607 [2024-10-01 20:22:15.778392] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:23:20.607 [2024-10-01 20:22:15.778424] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:23:20.866 [2024-10-01 20:22:16.045180] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:23:22.769 20:22:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:23:22.769 00:23:22.769 real 0m9.340s 00:23:22.769 user 0m14.601s 00:23:22.769 sys 0m1.361s 00:23:22.769 20:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:23:22.769 20:22:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:23:22.769 ************************************ 00:23:22.769 END TEST raid_superblock_test 00:23:22.769 ************************************ 00:23:22.769 20:22:17 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 3 read 00:23:22.769 20:22:17 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:23:22.769 20:22:17 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:23:22.769 20:22:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:23:22.769 ************************************ 00:23:22.769 START TEST raid_read_error_test 00:23:22.769 ************************************ 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 3 read 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.XIrHZPqSB2 00:23:22.769 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=69789 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 69789 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 69789 ']' 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:22.769 20:22:17 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:23.027 [2024-10-01 20:22:18.050254] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:23:23.027 [2024-10-01 20:22:18.050467] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69789 ] 00:23:23.027 [2024-10-01 20:22:18.226669] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:23.286 [2024-10-01 20:22:18.480982] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:23:23.545 [2024-10-01 20:22:18.683953] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:23:23.545 [2024-10-01 20:22:18.684059] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:24.112 BaseBdev1_malloc 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:24.112 true 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:24.112 [2024-10-01 20:22:19.151214] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:23:24.112 [2024-10-01 20:22:19.151291] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:24.112 [2024-10-01 20:22:19.151322] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:23:24.112 [2024-10-01 20:22:19.151342] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:24.112 [2024-10-01 20:22:19.154449] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:24.112 [2024-10-01 20:22:19.154500] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:23:24.112 BaseBdev1 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:24.112 BaseBdev2_malloc 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:24.112 true 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:24.112 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:24.112 [2024-10-01 20:22:19.210910] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:23:24.112 [2024-10-01 20:22:19.210995] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:24.112 [2024-10-01 20:22:19.211023] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:23:24.112 [2024-10-01 20:22:19.211041] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:24.112 [2024-10-01 20:22:19.213939] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:24.112 [2024-10-01 20:22:19.214000] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:23:24.113 BaseBdev2 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:24.113 BaseBdev3_malloc 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:24.113 true 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:24.113 [2024-10-01 20:22:19.270217] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:23:24.113 [2024-10-01 20:22:19.270300] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:24.113 [2024-10-01 20:22:19.270329] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:23:24.113 [2024-10-01 20:22:19.270347] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:24.113 [2024-10-01 20:22:19.273397] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:24.113 [2024-10-01 20:22:19.273446] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:23:24.113 BaseBdev3 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:24.113 [2024-10-01 20:22:19.282363] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:23:24.113 [2024-10-01 20:22:19.284864] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:23:24.113 [2024-10-01 20:22:19.285002] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:23:24.113 [2024-10-01 20:22:19.285297] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:23:24.113 [2024-10-01 20:22:19.285325] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:23:24.113 [2024-10-01 20:22:19.285650] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:23:24.113 [2024-10-01 20:22:19.285934] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:23:24.113 [2024-10-01 20:22:19.285964] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:23:24.113 [2024-10-01 20:22:19.286169] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:24.113 "name": "raid_bdev1", 00:23:24.113 "uuid": "86e45b24-6854-4018-a6ab-adc13afa57f5", 00:23:24.113 "strip_size_kb": 0, 00:23:24.113 "state": "online", 00:23:24.113 "raid_level": "raid1", 00:23:24.113 "superblock": true, 00:23:24.113 "num_base_bdevs": 3, 00:23:24.113 "num_base_bdevs_discovered": 3, 00:23:24.113 "num_base_bdevs_operational": 3, 00:23:24.113 "base_bdevs_list": [ 00:23:24.113 { 00:23:24.113 "name": "BaseBdev1", 00:23:24.113 "uuid": "22ac01cd-d225-5628-b445-ac66e690c04c", 00:23:24.113 "is_configured": true, 00:23:24.113 "data_offset": 2048, 00:23:24.113 "data_size": 63488 00:23:24.113 }, 00:23:24.113 { 00:23:24.113 "name": "BaseBdev2", 00:23:24.113 "uuid": "c8119878-0483-5a77-b497-a14ee3728347", 00:23:24.113 "is_configured": true, 00:23:24.113 "data_offset": 2048, 00:23:24.113 "data_size": 63488 00:23:24.113 }, 00:23:24.113 { 00:23:24.113 "name": "BaseBdev3", 00:23:24.113 "uuid": "c6f39d08-d217-5a05-8a45-f9df53685c27", 00:23:24.113 "is_configured": true, 00:23:24.113 "data_offset": 2048, 00:23:24.113 "data_size": 63488 00:23:24.113 } 00:23:24.113 ] 00:23:24.113 }' 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:24.113 20:22:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:24.682 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:23:24.682 20:22:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:23:24.941 [2024-10-01 20:22:19.948026] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:25.877 "name": "raid_bdev1", 00:23:25.877 "uuid": "86e45b24-6854-4018-a6ab-adc13afa57f5", 00:23:25.877 "strip_size_kb": 0, 00:23:25.877 "state": "online", 00:23:25.877 "raid_level": "raid1", 00:23:25.877 "superblock": true, 00:23:25.877 "num_base_bdevs": 3, 00:23:25.877 "num_base_bdevs_discovered": 3, 00:23:25.877 "num_base_bdevs_operational": 3, 00:23:25.877 "base_bdevs_list": [ 00:23:25.877 { 00:23:25.877 "name": "BaseBdev1", 00:23:25.877 "uuid": "22ac01cd-d225-5628-b445-ac66e690c04c", 00:23:25.877 "is_configured": true, 00:23:25.877 "data_offset": 2048, 00:23:25.877 "data_size": 63488 00:23:25.877 }, 00:23:25.877 { 00:23:25.877 "name": "BaseBdev2", 00:23:25.877 "uuid": "c8119878-0483-5a77-b497-a14ee3728347", 00:23:25.877 "is_configured": true, 00:23:25.877 "data_offset": 2048, 00:23:25.877 "data_size": 63488 00:23:25.877 }, 00:23:25.877 { 00:23:25.877 "name": "BaseBdev3", 00:23:25.877 "uuid": "c6f39d08-d217-5a05-8a45-f9df53685c27", 00:23:25.877 "is_configured": true, 00:23:25.877 "data_offset": 2048, 00:23:25.877 "data_size": 63488 00:23:25.877 } 00:23:25.877 ] 00:23:25.877 }' 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:25.877 20:22:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:26.136 20:22:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:23:26.136 20:22:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:26.136 20:22:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:26.136 [2024-10-01 20:22:21.357522] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:23:26.136 [2024-10-01 20:22:21.357564] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:23:26.136 [2024-10-01 20:22:21.361087] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:23:26.136 [2024-10-01 20:22:21.361159] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:26.136 [2024-10-01 20:22:21.361320] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:23:26.136 [2024-10-01 20:22:21.361348] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:23:26.136 { 00:23:26.136 "results": [ 00:23:26.136 { 00:23:26.136 "job": "raid_bdev1", 00:23:26.136 "core_mask": "0x1", 00:23:26.136 "workload": "randrw", 00:23:26.136 "percentage": 50, 00:23:26.136 "status": "finished", 00:23:26.136 "queue_depth": 1, 00:23:26.136 "io_size": 131072, 00:23:26.136 "runtime": 1.406895, 00:23:26.136 "iops": 8951.627520177411, 00:23:26.136 "mibps": 1118.9534400221764, 00:23:26.136 "io_failed": 0, 00:23:26.136 "io_timeout": 0, 00:23:26.136 "avg_latency_us": 106.76201943205278, 00:23:26.136 "min_latency_us": 42.589090909090906, 00:23:26.136 "max_latency_us": 1951.1854545454546 00:23:26.136 } 00:23:26.136 ], 00:23:26.136 "core_count": 1 00:23:26.136 } 00:23:26.136 20:22:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:26.136 20:22:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 69789 00:23:26.136 20:22:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 69789 ']' 00:23:26.136 20:22:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 69789 00:23:26.136 20:22:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:23:26.136 20:22:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:26.136 20:22:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69789 00:23:26.395 20:22:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:23:26.395 killing process with pid 69789 00:23:26.395 20:22:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:23:26.395 20:22:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69789' 00:23:26.395 20:22:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 69789 00:23:26.395 20:22:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 69789 00:23:26.395 [2024-10-01 20:22:21.395411] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:23:26.395 [2024-10-01 20:22:21.606487] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:23:28.299 20:22:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.XIrHZPqSB2 00:23:28.299 20:22:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:23:28.299 20:22:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:23:28.299 20:22:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:23:28.299 20:22:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:23:28.299 20:22:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:23:28.299 20:22:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:23:28.299 20:22:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:23:28.299 00:23:28.299 real 0m5.510s 00:23:28.299 user 0m6.557s 00:23:28.299 sys 0m0.726s 00:23:28.299 20:22:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:23:28.299 ************************************ 00:23:28.299 END TEST raid_read_error_test 00:23:28.299 ************************************ 00:23:28.299 20:22:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:28.299 20:22:23 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 3 write 00:23:28.299 20:22:23 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:23:28.299 20:22:23 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:23:28.299 20:22:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:23:28.299 ************************************ 00:23:28.299 START TEST raid_write_error_test 00:23:28.299 ************************************ 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 3 write 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.yOPioS0fJ6 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=69940 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 69940 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 69940 ']' 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:28.299 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:28.299 20:22:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:28.559 [2024-10-01 20:22:23.622527] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:23:28.559 [2024-10-01 20:22:23.622778] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69940 ] 00:23:28.559 [2024-10-01 20:22:23.804549] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:29.126 [2024-10-01 20:22:24.090044] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:23:29.126 [2024-10-01 20:22:24.301926] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:23:29.127 [2024-10-01 20:22:24.301989] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:29.694 BaseBdev1_malloc 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:29.694 true 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:29.694 [2024-10-01 20:22:24.781591] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:23:29.694 [2024-10-01 20:22:24.781667] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:29.694 [2024-10-01 20:22:24.781698] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:23:29.694 [2024-10-01 20:22:24.781748] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:29.694 [2024-10-01 20:22:24.785128] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:29.694 [2024-10-01 20:22:24.785179] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:23:29.694 BaseBdev1 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:29.694 BaseBdev2_malloc 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:29.694 true 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:29.694 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:29.694 [2024-10-01 20:22:24.851910] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:23:29.694 [2024-10-01 20:22:24.852010] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:29.694 [2024-10-01 20:22:24.852039] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:23:29.694 [2024-10-01 20:22:24.852057] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:29.694 [2024-10-01 20:22:24.855280] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:29.695 [2024-10-01 20:22:24.855341] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:23:29.695 BaseBdev2 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:29.695 BaseBdev3_malloc 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:29.695 true 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:29.695 [2024-10-01 20:22:24.918956] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:23:29.695 [2024-10-01 20:22:24.919038] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:23:29.695 [2024-10-01 20:22:24.919067] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:23:29.695 [2024-10-01 20:22:24.919086] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:23:29.695 [2024-10-01 20:22:24.922181] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:23:29.695 [2024-10-01 20:22:24.922225] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:23:29.695 BaseBdev3 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:29.695 [2024-10-01 20:22:24.927155] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:23:29.695 [2024-10-01 20:22:24.929913] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:23:29.695 [2024-10-01 20:22:24.930043] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:23:29.695 [2024-10-01 20:22:24.930355] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:23:29.695 [2024-10-01 20:22:24.930374] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:23:29.695 [2024-10-01 20:22:24.930681] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:23:29.695 [2024-10-01 20:22:24.930936] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:23:29.695 [2024-10-01 20:22:24.930967] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:23:29.695 [2024-10-01 20:22:24.931236] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:29.695 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:29.954 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:29.954 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:29.954 "name": "raid_bdev1", 00:23:29.954 "uuid": "e06ac053-eee8-46f4-9e14-7b87d28e89c5", 00:23:29.954 "strip_size_kb": 0, 00:23:29.954 "state": "online", 00:23:29.954 "raid_level": "raid1", 00:23:29.954 "superblock": true, 00:23:29.954 "num_base_bdevs": 3, 00:23:29.954 "num_base_bdevs_discovered": 3, 00:23:29.954 "num_base_bdevs_operational": 3, 00:23:29.954 "base_bdevs_list": [ 00:23:29.954 { 00:23:29.954 "name": "BaseBdev1", 00:23:29.954 "uuid": "cc5046b2-5408-58e2-8588-bbb64537c043", 00:23:29.954 "is_configured": true, 00:23:29.954 "data_offset": 2048, 00:23:29.954 "data_size": 63488 00:23:29.954 }, 00:23:29.954 { 00:23:29.954 "name": "BaseBdev2", 00:23:29.954 "uuid": "b3a64113-d92a-5b42-b390-97b84c7e6003", 00:23:29.954 "is_configured": true, 00:23:29.954 "data_offset": 2048, 00:23:29.954 "data_size": 63488 00:23:29.954 }, 00:23:29.954 { 00:23:29.954 "name": "BaseBdev3", 00:23:29.954 "uuid": "871a2b4c-9b99-5b5d-b45d-b406ce757e2e", 00:23:29.954 "is_configured": true, 00:23:29.954 "data_offset": 2048, 00:23:29.954 "data_size": 63488 00:23:29.954 } 00:23:29.954 ] 00:23:29.954 }' 00:23:29.954 20:22:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:29.954 20:22:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:30.214 20:22:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:23:30.214 20:22:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:23:30.473 [2024-10-01 20:22:25.533118] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:31.410 [2024-10-01 20:22:26.411996] bdev_raid.c:2272:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:23:31.410 [2024-10-01 20:22:26.412106] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:23:31.410 [2024-10-01 20:22:26.412422] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000005fb0 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:31.410 "name": "raid_bdev1", 00:23:31.410 "uuid": "e06ac053-eee8-46f4-9e14-7b87d28e89c5", 00:23:31.410 "strip_size_kb": 0, 00:23:31.410 "state": "online", 00:23:31.410 "raid_level": "raid1", 00:23:31.410 "superblock": true, 00:23:31.410 "num_base_bdevs": 3, 00:23:31.410 "num_base_bdevs_discovered": 2, 00:23:31.410 "num_base_bdevs_operational": 2, 00:23:31.410 "base_bdevs_list": [ 00:23:31.410 { 00:23:31.410 "name": null, 00:23:31.410 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:31.410 "is_configured": false, 00:23:31.410 "data_offset": 0, 00:23:31.410 "data_size": 63488 00:23:31.410 }, 00:23:31.410 { 00:23:31.410 "name": "BaseBdev2", 00:23:31.410 "uuid": "b3a64113-d92a-5b42-b390-97b84c7e6003", 00:23:31.410 "is_configured": true, 00:23:31.410 "data_offset": 2048, 00:23:31.410 "data_size": 63488 00:23:31.410 }, 00:23:31.410 { 00:23:31.410 "name": "BaseBdev3", 00:23:31.410 "uuid": "871a2b4c-9b99-5b5d-b45d-b406ce757e2e", 00:23:31.410 "is_configured": true, 00:23:31.410 "data_offset": 2048, 00:23:31.410 "data_size": 63488 00:23:31.410 } 00:23:31.410 ] 00:23:31.410 }' 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:31.410 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:31.978 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:23:31.978 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:31.978 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:31.978 [2024-10-01 20:22:26.949010] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:23:31.978 [2024-10-01 20:22:26.949066] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:23:31.978 [2024-10-01 20:22:26.952513] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:23:31.978 [2024-10-01 20:22:26.952593] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:31.978 [2024-10-01 20:22:26.952696] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:23:31.978 [2024-10-01 20:22:26.952741] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:23:31.978 { 00:23:31.978 "results": [ 00:23:31.978 { 00:23:31.978 "job": "raid_bdev1", 00:23:31.978 "core_mask": "0x1", 00:23:31.978 "workload": "randrw", 00:23:31.978 "percentage": 50, 00:23:31.978 "status": "finished", 00:23:31.978 "queue_depth": 1, 00:23:31.978 "io_size": 131072, 00:23:31.978 "runtime": 1.411903, 00:23:31.978 "iops": 9786.791302235351, 00:23:31.978 "mibps": 1223.348912779419, 00:23:31.978 "io_failed": 0, 00:23:31.978 "io_timeout": 0, 00:23:31.978 "avg_latency_us": 97.53014223871367, 00:23:31.978 "min_latency_us": 41.192727272727275, 00:23:31.978 "max_latency_us": 1832.0290909090909 00:23:31.978 } 00:23:31.978 ], 00:23:31.978 "core_count": 1 00:23:31.978 } 00:23:31.978 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:31.978 20:22:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 69940 00:23:31.978 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 69940 ']' 00:23:31.978 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 69940 00:23:31.978 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:23:31.978 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:31.978 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69940 00:23:31.978 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:23:31.978 killing process with pid 69940 00:23:31.978 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:23:31.978 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69940' 00:23:31.978 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 69940 00:23:31.978 [2024-10-01 20:22:26.988617] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:23:31.978 20:22:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 69940 00:23:31.978 [2024-10-01 20:22:27.205555] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:23:33.911 20:22:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.yOPioS0fJ6 00:23:33.911 20:22:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:23:33.911 20:22:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:23:33.911 20:22:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:23:33.911 20:22:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:23:33.911 20:22:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:23:33.911 20:22:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:23:33.911 20:22:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:23:33.911 00:23:33.911 real 0m5.556s 00:23:33.911 user 0m6.525s 00:23:33.911 sys 0m0.721s 00:23:33.911 20:22:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:23:33.911 20:22:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:23:33.911 ************************************ 00:23:33.911 END TEST raid_write_error_test 00:23:33.911 ************************************ 00:23:33.911 20:22:29 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:23:33.911 20:22:29 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:23:33.911 20:22:29 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 4 false 00:23:33.911 20:22:29 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:23:33.911 20:22:29 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:23:33.911 20:22:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:23:33.911 ************************************ 00:23:33.911 START TEST raid_state_function_test 00:23:33.911 ************************************ 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 4 false 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=70095 00:23:33.911 Process raid pid: 70095 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 70095' 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 70095 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 70095 ']' 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:33.911 20:22:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:33.911 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:33.912 20:22:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:33.912 20:22:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:33.912 20:22:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:34.171 [2024-10-01 20:22:29.228357] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:23:34.171 [2024-10-01 20:22:29.228557] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:34.171 [2024-10-01 20:22:29.407181] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:34.429 [2024-10-01 20:22:29.649962] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:23:34.737 [2024-10-01 20:22:29.855417] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:23:34.737 [2024-10-01 20:22:29.855506] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:35.305 [2024-10-01 20:22:30.284468] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:23:35.305 [2024-10-01 20:22:30.284549] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:23:35.305 [2024-10-01 20:22:30.284565] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:23:35.305 [2024-10-01 20:22:30.284614] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:23:35.305 [2024-10-01 20:22:30.284625] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:23:35.305 [2024-10-01 20:22:30.284643] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:23:35.305 [2024-10-01 20:22:30.284654] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:23:35.305 [2024-10-01 20:22:30.284669] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:35.305 "name": "Existed_Raid", 00:23:35.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:35.305 "strip_size_kb": 64, 00:23:35.305 "state": "configuring", 00:23:35.305 "raid_level": "raid0", 00:23:35.305 "superblock": false, 00:23:35.305 "num_base_bdevs": 4, 00:23:35.305 "num_base_bdevs_discovered": 0, 00:23:35.305 "num_base_bdevs_operational": 4, 00:23:35.305 "base_bdevs_list": [ 00:23:35.305 { 00:23:35.305 "name": "BaseBdev1", 00:23:35.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:35.305 "is_configured": false, 00:23:35.305 "data_offset": 0, 00:23:35.305 "data_size": 0 00:23:35.305 }, 00:23:35.305 { 00:23:35.305 "name": "BaseBdev2", 00:23:35.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:35.305 "is_configured": false, 00:23:35.305 "data_offset": 0, 00:23:35.305 "data_size": 0 00:23:35.305 }, 00:23:35.305 { 00:23:35.305 "name": "BaseBdev3", 00:23:35.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:35.305 "is_configured": false, 00:23:35.305 "data_offset": 0, 00:23:35.305 "data_size": 0 00:23:35.305 }, 00:23:35.305 { 00:23:35.305 "name": "BaseBdev4", 00:23:35.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:35.305 "is_configured": false, 00:23:35.305 "data_offset": 0, 00:23:35.305 "data_size": 0 00:23:35.305 } 00:23:35.305 ] 00:23:35.305 }' 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:35.305 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:35.564 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:23:35.564 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:35.564 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:35.564 [2024-10-01 20:22:30.752517] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:23:35.564 [2024-10-01 20:22:30.752584] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:23:35.564 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:35.564 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:23:35.564 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:35.564 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:35.564 [2024-10-01 20:22:30.760516] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:23:35.564 [2024-10-01 20:22:30.760581] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:23:35.564 [2024-10-01 20:22:30.760596] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:23:35.564 [2024-10-01 20:22:30.760613] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:23:35.564 [2024-10-01 20:22:30.760623] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:23:35.564 [2024-10-01 20:22:30.760637] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:23:35.564 [2024-10-01 20:22:30.760653] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:23:35.565 [2024-10-01 20:22:30.760668] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:23:35.565 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:35.565 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:23:35.565 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:35.565 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:35.565 [2024-10-01 20:22:30.806782] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:23:35.565 BaseBdev1 00:23:35.565 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:35.565 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:23:35.565 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:23:35.565 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:35.565 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:23:35.565 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:35.565 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:35.565 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:35.565 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:35.565 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:35.824 [ 00:23:35.824 { 00:23:35.824 "name": "BaseBdev1", 00:23:35.824 "aliases": [ 00:23:35.824 "adb510ec-2280-46ea-8f76-053b24a560c3" 00:23:35.824 ], 00:23:35.824 "product_name": "Malloc disk", 00:23:35.824 "block_size": 512, 00:23:35.824 "num_blocks": 65536, 00:23:35.824 "uuid": "adb510ec-2280-46ea-8f76-053b24a560c3", 00:23:35.824 "assigned_rate_limits": { 00:23:35.824 "rw_ios_per_sec": 0, 00:23:35.824 "rw_mbytes_per_sec": 0, 00:23:35.824 "r_mbytes_per_sec": 0, 00:23:35.824 "w_mbytes_per_sec": 0 00:23:35.824 }, 00:23:35.824 "claimed": true, 00:23:35.824 "claim_type": "exclusive_write", 00:23:35.824 "zoned": false, 00:23:35.824 "supported_io_types": { 00:23:35.824 "read": true, 00:23:35.824 "write": true, 00:23:35.824 "unmap": true, 00:23:35.824 "flush": true, 00:23:35.824 "reset": true, 00:23:35.824 "nvme_admin": false, 00:23:35.824 "nvme_io": false, 00:23:35.824 "nvme_io_md": false, 00:23:35.824 "write_zeroes": true, 00:23:35.824 "zcopy": true, 00:23:35.824 "get_zone_info": false, 00:23:35.824 "zone_management": false, 00:23:35.824 "zone_append": false, 00:23:35.824 "compare": false, 00:23:35.824 "compare_and_write": false, 00:23:35.824 "abort": true, 00:23:35.824 "seek_hole": false, 00:23:35.824 "seek_data": false, 00:23:35.824 "copy": true, 00:23:35.824 "nvme_iov_md": false 00:23:35.824 }, 00:23:35.824 "memory_domains": [ 00:23:35.824 { 00:23:35.824 "dma_device_id": "system", 00:23:35.824 "dma_device_type": 1 00:23:35.824 }, 00:23:35.824 { 00:23:35.824 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:35.824 "dma_device_type": 2 00:23:35.824 } 00:23:35.824 ], 00:23:35.824 "driver_specific": {} 00:23:35.824 } 00:23:35.824 ] 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:35.824 "name": "Existed_Raid", 00:23:35.824 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:35.824 "strip_size_kb": 64, 00:23:35.824 "state": "configuring", 00:23:35.824 "raid_level": "raid0", 00:23:35.824 "superblock": false, 00:23:35.824 "num_base_bdevs": 4, 00:23:35.824 "num_base_bdevs_discovered": 1, 00:23:35.824 "num_base_bdevs_operational": 4, 00:23:35.824 "base_bdevs_list": [ 00:23:35.824 { 00:23:35.824 "name": "BaseBdev1", 00:23:35.824 "uuid": "adb510ec-2280-46ea-8f76-053b24a560c3", 00:23:35.824 "is_configured": true, 00:23:35.824 "data_offset": 0, 00:23:35.824 "data_size": 65536 00:23:35.824 }, 00:23:35.824 { 00:23:35.824 "name": "BaseBdev2", 00:23:35.824 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:35.824 "is_configured": false, 00:23:35.824 "data_offset": 0, 00:23:35.824 "data_size": 0 00:23:35.824 }, 00:23:35.824 { 00:23:35.824 "name": "BaseBdev3", 00:23:35.824 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:35.824 "is_configured": false, 00:23:35.824 "data_offset": 0, 00:23:35.824 "data_size": 0 00:23:35.824 }, 00:23:35.824 { 00:23:35.824 "name": "BaseBdev4", 00:23:35.824 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:35.824 "is_configured": false, 00:23:35.824 "data_offset": 0, 00:23:35.824 "data_size": 0 00:23:35.824 } 00:23:35.824 ] 00:23:35.824 }' 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:35.824 20:22:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:36.082 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:23:36.082 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:36.082 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:36.082 [2024-10-01 20:22:31.331027] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:23:36.082 [2024-10-01 20:22:31.331158] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:23:36.339 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:36.339 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:23:36.339 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:36.339 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:36.339 [2024-10-01 20:22:31.339100] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:23:36.340 [2024-10-01 20:22:31.342452] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:23:36.340 [2024-10-01 20:22:31.342528] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:23:36.340 [2024-10-01 20:22:31.342548] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:23:36.340 [2024-10-01 20:22:31.342570] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:23:36.340 [2024-10-01 20:22:31.342591] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:23:36.340 [2024-10-01 20:22:31.342609] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:36.340 "name": "Existed_Raid", 00:23:36.340 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:36.340 "strip_size_kb": 64, 00:23:36.340 "state": "configuring", 00:23:36.340 "raid_level": "raid0", 00:23:36.340 "superblock": false, 00:23:36.340 "num_base_bdevs": 4, 00:23:36.340 "num_base_bdevs_discovered": 1, 00:23:36.340 "num_base_bdevs_operational": 4, 00:23:36.340 "base_bdevs_list": [ 00:23:36.340 { 00:23:36.340 "name": "BaseBdev1", 00:23:36.340 "uuid": "adb510ec-2280-46ea-8f76-053b24a560c3", 00:23:36.340 "is_configured": true, 00:23:36.340 "data_offset": 0, 00:23:36.340 "data_size": 65536 00:23:36.340 }, 00:23:36.340 { 00:23:36.340 "name": "BaseBdev2", 00:23:36.340 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:36.340 "is_configured": false, 00:23:36.340 "data_offset": 0, 00:23:36.340 "data_size": 0 00:23:36.340 }, 00:23:36.340 { 00:23:36.340 "name": "BaseBdev3", 00:23:36.340 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:36.340 "is_configured": false, 00:23:36.340 "data_offset": 0, 00:23:36.340 "data_size": 0 00:23:36.340 }, 00:23:36.340 { 00:23:36.340 "name": "BaseBdev4", 00:23:36.340 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:36.340 "is_configured": false, 00:23:36.340 "data_offset": 0, 00:23:36.340 "data_size": 0 00:23:36.340 } 00:23:36.340 ] 00:23:36.340 }' 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:36.340 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:36.907 [2024-10-01 20:22:31.901165] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:23:36.907 BaseBdev2 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:36.907 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:36.907 [ 00:23:36.907 { 00:23:36.907 "name": "BaseBdev2", 00:23:36.907 "aliases": [ 00:23:36.907 "6314415d-4004-4b73-8cf6-4d5fa7d495a0" 00:23:36.907 ], 00:23:36.907 "product_name": "Malloc disk", 00:23:36.907 "block_size": 512, 00:23:36.907 "num_blocks": 65536, 00:23:36.907 "uuid": "6314415d-4004-4b73-8cf6-4d5fa7d495a0", 00:23:36.907 "assigned_rate_limits": { 00:23:36.907 "rw_ios_per_sec": 0, 00:23:36.907 "rw_mbytes_per_sec": 0, 00:23:36.907 "r_mbytes_per_sec": 0, 00:23:36.907 "w_mbytes_per_sec": 0 00:23:36.907 }, 00:23:36.907 "claimed": true, 00:23:36.907 "claim_type": "exclusive_write", 00:23:36.907 "zoned": false, 00:23:36.907 "supported_io_types": { 00:23:36.907 "read": true, 00:23:36.907 "write": true, 00:23:36.907 "unmap": true, 00:23:36.907 "flush": true, 00:23:36.907 "reset": true, 00:23:36.907 "nvme_admin": false, 00:23:36.907 "nvme_io": false, 00:23:36.907 "nvme_io_md": false, 00:23:36.907 "write_zeroes": true, 00:23:36.907 "zcopy": true, 00:23:36.907 "get_zone_info": false, 00:23:36.907 "zone_management": false, 00:23:36.907 "zone_append": false, 00:23:36.907 "compare": false, 00:23:36.907 "compare_and_write": false, 00:23:36.907 "abort": true, 00:23:36.907 "seek_hole": false, 00:23:36.907 "seek_data": false, 00:23:36.907 "copy": true, 00:23:36.907 "nvme_iov_md": false 00:23:36.907 }, 00:23:36.907 "memory_domains": [ 00:23:36.907 { 00:23:36.907 "dma_device_id": "system", 00:23:36.907 "dma_device_type": 1 00:23:36.907 }, 00:23:36.908 { 00:23:36.908 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:36.908 "dma_device_type": 2 00:23:36.908 } 00:23:36.908 ], 00:23:36.908 "driver_specific": {} 00:23:36.908 } 00:23:36.908 ] 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:36.908 "name": "Existed_Raid", 00:23:36.908 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:36.908 "strip_size_kb": 64, 00:23:36.908 "state": "configuring", 00:23:36.908 "raid_level": "raid0", 00:23:36.908 "superblock": false, 00:23:36.908 "num_base_bdevs": 4, 00:23:36.908 "num_base_bdevs_discovered": 2, 00:23:36.908 "num_base_bdevs_operational": 4, 00:23:36.908 "base_bdevs_list": [ 00:23:36.908 { 00:23:36.908 "name": "BaseBdev1", 00:23:36.908 "uuid": "adb510ec-2280-46ea-8f76-053b24a560c3", 00:23:36.908 "is_configured": true, 00:23:36.908 "data_offset": 0, 00:23:36.908 "data_size": 65536 00:23:36.908 }, 00:23:36.908 { 00:23:36.908 "name": "BaseBdev2", 00:23:36.908 "uuid": "6314415d-4004-4b73-8cf6-4d5fa7d495a0", 00:23:36.908 "is_configured": true, 00:23:36.908 "data_offset": 0, 00:23:36.908 "data_size": 65536 00:23:36.908 }, 00:23:36.908 { 00:23:36.908 "name": "BaseBdev3", 00:23:36.908 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:36.908 "is_configured": false, 00:23:36.908 "data_offset": 0, 00:23:36.908 "data_size": 0 00:23:36.908 }, 00:23:36.908 { 00:23:36.908 "name": "BaseBdev4", 00:23:36.908 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:36.908 "is_configured": false, 00:23:36.908 "data_offset": 0, 00:23:36.908 "data_size": 0 00:23:36.908 } 00:23:36.908 ] 00:23:36.908 }' 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:36.908 20:22:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:37.474 [2024-10-01 20:22:32.517066] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:23:37.474 BaseBdev3 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:37.474 [ 00:23:37.474 { 00:23:37.474 "name": "BaseBdev3", 00:23:37.474 "aliases": [ 00:23:37.474 "fd7f8358-94c9-4447-99ec-d6760095a2e3" 00:23:37.474 ], 00:23:37.474 "product_name": "Malloc disk", 00:23:37.474 "block_size": 512, 00:23:37.474 "num_blocks": 65536, 00:23:37.474 "uuid": "fd7f8358-94c9-4447-99ec-d6760095a2e3", 00:23:37.474 "assigned_rate_limits": { 00:23:37.474 "rw_ios_per_sec": 0, 00:23:37.474 "rw_mbytes_per_sec": 0, 00:23:37.474 "r_mbytes_per_sec": 0, 00:23:37.474 "w_mbytes_per_sec": 0 00:23:37.474 }, 00:23:37.474 "claimed": true, 00:23:37.474 "claim_type": "exclusive_write", 00:23:37.474 "zoned": false, 00:23:37.474 "supported_io_types": { 00:23:37.474 "read": true, 00:23:37.474 "write": true, 00:23:37.474 "unmap": true, 00:23:37.474 "flush": true, 00:23:37.474 "reset": true, 00:23:37.474 "nvme_admin": false, 00:23:37.474 "nvme_io": false, 00:23:37.474 "nvme_io_md": false, 00:23:37.474 "write_zeroes": true, 00:23:37.474 "zcopy": true, 00:23:37.474 "get_zone_info": false, 00:23:37.474 "zone_management": false, 00:23:37.474 "zone_append": false, 00:23:37.474 "compare": false, 00:23:37.474 "compare_and_write": false, 00:23:37.474 "abort": true, 00:23:37.474 "seek_hole": false, 00:23:37.474 "seek_data": false, 00:23:37.474 "copy": true, 00:23:37.474 "nvme_iov_md": false 00:23:37.474 }, 00:23:37.474 "memory_domains": [ 00:23:37.474 { 00:23:37.474 "dma_device_id": "system", 00:23:37.474 "dma_device_type": 1 00:23:37.474 }, 00:23:37.474 { 00:23:37.474 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:37.474 "dma_device_type": 2 00:23:37.474 } 00:23:37.474 ], 00:23:37.474 "driver_specific": {} 00:23:37.474 } 00:23:37.474 ] 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:37.474 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:37.474 "name": "Existed_Raid", 00:23:37.474 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:37.474 "strip_size_kb": 64, 00:23:37.474 "state": "configuring", 00:23:37.474 "raid_level": "raid0", 00:23:37.474 "superblock": false, 00:23:37.474 "num_base_bdevs": 4, 00:23:37.474 "num_base_bdevs_discovered": 3, 00:23:37.474 "num_base_bdevs_operational": 4, 00:23:37.474 "base_bdevs_list": [ 00:23:37.474 { 00:23:37.474 "name": "BaseBdev1", 00:23:37.474 "uuid": "adb510ec-2280-46ea-8f76-053b24a560c3", 00:23:37.474 "is_configured": true, 00:23:37.474 "data_offset": 0, 00:23:37.474 "data_size": 65536 00:23:37.474 }, 00:23:37.474 { 00:23:37.474 "name": "BaseBdev2", 00:23:37.474 "uuid": "6314415d-4004-4b73-8cf6-4d5fa7d495a0", 00:23:37.474 "is_configured": true, 00:23:37.474 "data_offset": 0, 00:23:37.474 "data_size": 65536 00:23:37.474 }, 00:23:37.474 { 00:23:37.474 "name": "BaseBdev3", 00:23:37.474 "uuid": "fd7f8358-94c9-4447-99ec-d6760095a2e3", 00:23:37.474 "is_configured": true, 00:23:37.475 "data_offset": 0, 00:23:37.475 "data_size": 65536 00:23:37.475 }, 00:23:37.475 { 00:23:37.475 "name": "BaseBdev4", 00:23:37.475 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:37.475 "is_configured": false, 00:23:37.475 "data_offset": 0, 00:23:37.475 "data_size": 0 00:23:37.475 } 00:23:37.475 ] 00:23:37.475 }' 00:23:37.475 20:22:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:37.475 20:22:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:38.041 [2024-10-01 20:22:33.107317] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:23:38.041 [2024-10-01 20:22:33.107392] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:23:38.041 [2024-10-01 20:22:33.107407] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:23:38.041 [2024-10-01 20:22:33.107816] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:23:38.041 [2024-10-01 20:22:33.108044] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:23:38.041 [2024-10-01 20:22:33.108080] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:23:38.041 [2024-10-01 20:22:33.108402] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:38.041 BaseBdev4 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:38.041 [ 00:23:38.041 { 00:23:38.041 "name": "BaseBdev4", 00:23:38.041 "aliases": [ 00:23:38.041 "9d8957f0-bd1f-4864-93aa-e82744efa080" 00:23:38.041 ], 00:23:38.041 "product_name": "Malloc disk", 00:23:38.041 "block_size": 512, 00:23:38.041 "num_blocks": 65536, 00:23:38.041 "uuid": "9d8957f0-bd1f-4864-93aa-e82744efa080", 00:23:38.041 "assigned_rate_limits": { 00:23:38.041 "rw_ios_per_sec": 0, 00:23:38.041 "rw_mbytes_per_sec": 0, 00:23:38.041 "r_mbytes_per_sec": 0, 00:23:38.041 "w_mbytes_per_sec": 0 00:23:38.041 }, 00:23:38.041 "claimed": true, 00:23:38.041 "claim_type": "exclusive_write", 00:23:38.041 "zoned": false, 00:23:38.041 "supported_io_types": { 00:23:38.041 "read": true, 00:23:38.041 "write": true, 00:23:38.041 "unmap": true, 00:23:38.041 "flush": true, 00:23:38.041 "reset": true, 00:23:38.041 "nvme_admin": false, 00:23:38.041 "nvme_io": false, 00:23:38.041 "nvme_io_md": false, 00:23:38.041 "write_zeroes": true, 00:23:38.041 "zcopy": true, 00:23:38.041 "get_zone_info": false, 00:23:38.041 "zone_management": false, 00:23:38.041 "zone_append": false, 00:23:38.041 "compare": false, 00:23:38.041 "compare_and_write": false, 00:23:38.041 "abort": true, 00:23:38.041 "seek_hole": false, 00:23:38.041 "seek_data": false, 00:23:38.041 "copy": true, 00:23:38.041 "nvme_iov_md": false 00:23:38.041 }, 00:23:38.041 "memory_domains": [ 00:23:38.041 { 00:23:38.041 "dma_device_id": "system", 00:23:38.041 "dma_device_type": 1 00:23:38.041 }, 00:23:38.041 { 00:23:38.041 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:38.041 "dma_device_type": 2 00:23:38.041 } 00:23:38.041 ], 00:23:38.041 "driver_specific": {} 00:23:38.041 } 00:23:38.041 ] 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:38.041 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:38.041 "name": "Existed_Raid", 00:23:38.041 "uuid": "aa163c75-c8be-4750-acb7-ea8815e3ede9", 00:23:38.041 "strip_size_kb": 64, 00:23:38.041 "state": "online", 00:23:38.041 "raid_level": "raid0", 00:23:38.041 "superblock": false, 00:23:38.041 "num_base_bdevs": 4, 00:23:38.041 "num_base_bdevs_discovered": 4, 00:23:38.041 "num_base_bdevs_operational": 4, 00:23:38.041 "base_bdevs_list": [ 00:23:38.041 { 00:23:38.041 "name": "BaseBdev1", 00:23:38.041 "uuid": "adb510ec-2280-46ea-8f76-053b24a560c3", 00:23:38.041 "is_configured": true, 00:23:38.041 "data_offset": 0, 00:23:38.041 "data_size": 65536 00:23:38.041 }, 00:23:38.041 { 00:23:38.041 "name": "BaseBdev2", 00:23:38.041 "uuid": "6314415d-4004-4b73-8cf6-4d5fa7d495a0", 00:23:38.041 "is_configured": true, 00:23:38.041 "data_offset": 0, 00:23:38.041 "data_size": 65536 00:23:38.041 }, 00:23:38.041 { 00:23:38.041 "name": "BaseBdev3", 00:23:38.041 "uuid": "fd7f8358-94c9-4447-99ec-d6760095a2e3", 00:23:38.041 "is_configured": true, 00:23:38.042 "data_offset": 0, 00:23:38.042 "data_size": 65536 00:23:38.042 }, 00:23:38.042 { 00:23:38.042 "name": "BaseBdev4", 00:23:38.042 "uuid": "9d8957f0-bd1f-4864-93aa-e82744efa080", 00:23:38.042 "is_configured": true, 00:23:38.042 "data_offset": 0, 00:23:38.042 "data_size": 65536 00:23:38.042 } 00:23:38.042 ] 00:23:38.042 }' 00:23:38.042 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:38.042 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:38.607 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:23:38.607 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:23:38.607 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:23:38.607 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:23:38.607 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:23:38.607 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:23:38.607 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:23:38.607 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:38.607 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:38.607 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:23:38.607 [2024-10-01 20:22:33.652015] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:23:38.607 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:38.607 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:38.607 "name": "Existed_Raid", 00:23:38.607 "aliases": [ 00:23:38.607 "aa163c75-c8be-4750-acb7-ea8815e3ede9" 00:23:38.607 ], 00:23:38.607 "product_name": "Raid Volume", 00:23:38.607 "block_size": 512, 00:23:38.607 "num_blocks": 262144, 00:23:38.607 "uuid": "aa163c75-c8be-4750-acb7-ea8815e3ede9", 00:23:38.607 "assigned_rate_limits": { 00:23:38.607 "rw_ios_per_sec": 0, 00:23:38.607 "rw_mbytes_per_sec": 0, 00:23:38.607 "r_mbytes_per_sec": 0, 00:23:38.607 "w_mbytes_per_sec": 0 00:23:38.607 }, 00:23:38.607 "claimed": false, 00:23:38.607 "zoned": false, 00:23:38.607 "supported_io_types": { 00:23:38.607 "read": true, 00:23:38.607 "write": true, 00:23:38.607 "unmap": true, 00:23:38.607 "flush": true, 00:23:38.607 "reset": true, 00:23:38.607 "nvme_admin": false, 00:23:38.607 "nvme_io": false, 00:23:38.607 "nvme_io_md": false, 00:23:38.607 "write_zeroes": true, 00:23:38.607 "zcopy": false, 00:23:38.607 "get_zone_info": false, 00:23:38.607 "zone_management": false, 00:23:38.607 "zone_append": false, 00:23:38.607 "compare": false, 00:23:38.607 "compare_and_write": false, 00:23:38.607 "abort": false, 00:23:38.607 "seek_hole": false, 00:23:38.607 "seek_data": false, 00:23:38.607 "copy": false, 00:23:38.607 "nvme_iov_md": false 00:23:38.607 }, 00:23:38.607 "memory_domains": [ 00:23:38.607 { 00:23:38.607 "dma_device_id": "system", 00:23:38.607 "dma_device_type": 1 00:23:38.607 }, 00:23:38.607 { 00:23:38.607 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:38.607 "dma_device_type": 2 00:23:38.607 }, 00:23:38.607 { 00:23:38.607 "dma_device_id": "system", 00:23:38.607 "dma_device_type": 1 00:23:38.607 }, 00:23:38.607 { 00:23:38.607 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:38.607 "dma_device_type": 2 00:23:38.607 }, 00:23:38.607 { 00:23:38.607 "dma_device_id": "system", 00:23:38.607 "dma_device_type": 1 00:23:38.607 }, 00:23:38.607 { 00:23:38.607 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:38.607 "dma_device_type": 2 00:23:38.607 }, 00:23:38.607 { 00:23:38.607 "dma_device_id": "system", 00:23:38.607 "dma_device_type": 1 00:23:38.607 }, 00:23:38.607 { 00:23:38.607 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:38.607 "dma_device_type": 2 00:23:38.607 } 00:23:38.607 ], 00:23:38.607 "driver_specific": { 00:23:38.607 "raid": { 00:23:38.607 "uuid": "aa163c75-c8be-4750-acb7-ea8815e3ede9", 00:23:38.607 "strip_size_kb": 64, 00:23:38.607 "state": "online", 00:23:38.607 "raid_level": "raid0", 00:23:38.607 "superblock": false, 00:23:38.607 "num_base_bdevs": 4, 00:23:38.607 "num_base_bdevs_discovered": 4, 00:23:38.607 "num_base_bdevs_operational": 4, 00:23:38.607 "base_bdevs_list": [ 00:23:38.607 { 00:23:38.607 "name": "BaseBdev1", 00:23:38.607 "uuid": "adb510ec-2280-46ea-8f76-053b24a560c3", 00:23:38.607 "is_configured": true, 00:23:38.607 "data_offset": 0, 00:23:38.607 "data_size": 65536 00:23:38.607 }, 00:23:38.607 { 00:23:38.607 "name": "BaseBdev2", 00:23:38.607 "uuid": "6314415d-4004-4b73-8cf6-4d5fa7d495a0", 00:23:38.607 "is_configured": true, 00:23:38.607 "data_offset": 0, 00:23:38.607 "data_size": 65536 00:23:38.607 }, 00:23:38.607 { 00:23:38.607 "name": "BaseBdev3", 00:23:38.607 "uuid": "fd7f8358-94c9-4447-99ec-d6760095a2e3", 00:23:38.607 "is_configured": true, 00:23:38.607 "data_offset": 0, 00:23:38.607 "data_size": 65536 00:23:38.607 }, 00:23:38.607 { 00:23:38.608 "name": "BaseBdev4", 00:23:38.608 "uuid": "9d8957f0-bd1f-4864-93aa-e82744efa080", 00:23:38.608 "is_configured": true, 00:23:38.608 "data_offset": 0, 00:23:38.608 "data_size": 65536 00:23:38.608 } 00:23:38.608 ] 00:23:38.608 } 00:23:38.608 } 00:23:38.608 }' 00:23:38.608 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:23:38.608 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:23:38.608 BaseBdev2 00:23:38.608 BaseBdev3 00:23:38.608 BaseBdev4' 00:23:38.608 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:38.608 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:23:38.608 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:38.608 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:23:38.608 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:38.608 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:38.608 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:38.608 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:38.866 20:22:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:38.866 [2024-10-01 20:22:34.023706] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:23:38.866 [2024-10-01 20:22:34.023796] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:23:38.866 [2024-10-01 20:22:34.023872] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:38.866 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:39.124 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:39.124 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:39.124 "name": "Existed_Raid", 00:23:39.124 "uuid": "aa163c75-c8be-4750-acb7-ea8815e3ede9", 00:23:39.124 "strip_size_kb": 64, 00:23:39.124 "state": "offline", 00:23:39.124 "raid_level": "raid0", 00:23:39.124 "superblock": false, 00:23:39.124 "num_base_bdevs": 4, 00:23:39.124 "num_base_bdevs_discovered": 3, 00:23:39.124 "num_base_bdevs_operational": 3, 00:23:39.124 "base_bdevs_list": [ 00:23:39.124 { 00:23:39.124 "name": null, 00:23:39.124 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:39.124 "is_configured": false, 00:23:39.124 "data_offset": 0, 00:23:39.124 "data_size": 65536 00:23:39.124 }, 00:23:39.124 { 00:23:39.124 "name": "BaseBdev2", 00:23:39.124 "uuid": "6314415d-4004-4b73-8cf6-4d5fa7d495a0", 00:23:39.124 "is_configured": true, 00:23:39.124 "data_offset": 0, 00:23:39.124 "data_size": 65536 00:23:39.124 }, 00:23:39.124 { 00:23:39.124 "name": "BaseBdev3", 00:23:39.124 "uuid": "fd7f8358-94c9-4447-99ec-d6760095a2e3", 00:23:39.124 "is_configured": true, 00:23:39.124 "data_offset": 0, 00:23:39.124 "data_size": 65536 00:23:39.124 }, 00:23:39.124 { 00:23:39.124 "name": "BaseBdev4", 00:23:39.124 "uuid": "9d8957f0-bd1f-4864-93aa-e82744efa080", 00:23:39.124 "is_configured": true, 00:23:39.124 "data_offset": 0, 00:23:39.124 "data_size": 65536 00:23:39.124 } 00:23:39.124 ] 00:23:39.124 }' 00:23:39.124 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:39.124 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:39.382 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:23:39.382 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:23:39.382 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:39.382 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:39.382 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:23:39.382 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:39.639 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:39.639 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:23:39.639 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:23:39.639 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:23:39.639 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:39.640 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:39.640 [2024-10-01 20:22:34.678484] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:23:39.640 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:39.640 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:23:39.640 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:23:39.640 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:39.640 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:39.640 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:23:39.640 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:39.640 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:39.640 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:23:39.640 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:23:39.640 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:23:39.640 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:39.640 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:39.640 [2024-10-01 20:22:34.818540] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:23:39.897 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:39.897 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:23:39.898 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:23:39.898 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:39.898 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:39.898 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:39.898 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:23:39.898 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:39.898 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:23:39.898 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:23:39.898 20:22:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:23:39.898 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:39.898 20:22:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:39.898 [2024-10-01 20:22:34.962547] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:23:39.898 [2024-10-01 20:22:34.962770] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:39.898 BaseBdev2 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:39.898 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:40.156 [ 00:23:40.156 { 00:23:40.156 "name": "BaseBdev2", 00:23:40.156 "aliases": [ 00:23:40.156 "d4af6e2c-79ce-417d-8b5f-476b5d3b4690" 00:23:40.156 ], 00:23:40.156 "product_name": "Malloc disk", 00:23:40.156 "block_size": 512, 00:23:40.156 "num_blocks": 65536, 00:23:40.156 "uuid": "d4af6e2c-79ce-417d-8b5f-476b5d3b4690", 00:23:40.156 "assigned_rate_limits": { 00:23:40.156 "rw_ios_per_sec": 0, 00:23:40.156 "rw_mbytes_per_sec": 0, 00:23:40.156 "r_mbytes_per_sec": 0, 00:23:40.156 "w_mbytes_per_sec": 0 00:23:40.156 }, 00:23:40.156 "claimed": false, 00:23:40.156 "zoned": false, 00:23:40.156 "supported_io_types": { 00:23:40.156 "read": true, 00:23:40.156 "write": true, 00:23:40.156 "unmap": true, 00:23:40.156 "flush": true, 00:23:40.156 "reset": true, 00:23:40.156 "nvme_admin": false, 00:23:40.156 "nvme_io": false, 00:23:40.156 "nvme_io_md": false, 00:23:40.156 "write_zeroes": true, 00:23:40.156 "zcopy": true, 00:23:40.156 "get_zone_info": false, 00:23:40.156 "zone_management": false, 00:23:40.156 "zone_append": false, 00:23:40.156 "compare": false, 00:23:40.156 "compare_and_write": false, 00:23:40.156 "abort": true, 00:23:40.156 "seek_hole": false, 00:23:40.156 "seek_data": false, 00:23:40.156 "copy": true, 00:23:40.156 "nvme_iov_md": false 00:23:40.156 }, 00:23:40.156 "memory_domains": [ 00:23:40.156 { 00:23:40.156 "dma_device_id": "system", 00:23:40.156 "dma_device_type": 1 00:23:40.156 }, 00:23:40.156 { 00:23:40.156 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:40.156 "dma_device_type": 2 00:23:40.156 } 00:23:40.156 ], 00:23:40.156 "driver_specific": {} 00:23:40.156 } 00:23:40.156 ] 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:40.156 BaseBdev3 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:23:40.156 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:40.157 [ 00:23:40.157 { 00:23:40.157 "name": "BaseBdev3", 00:23:40.157 "aliases": [ 00:23:40.157 "be32d2e4-fe7e-4908-b551-c3301540ed03" 00:23:40.157 ], 00:23:40.157 "product_name": "Malloc disk", 00:23:40.157 "block_size": 512, 00:23:40.157 "num_blocks": 65536, 00:23:40.157 "uuid": "be32d2e4-fe7e-4908-b551-c3301540ed03", 00:23:40.157 "assigned_rate_limits": { 00:23:40.157 "rw_ios_per_sec": 0, 00:23:40.157 "rw_mbytes_per_sec": 0, 00:23:40.157 "r_mbytes_per_sec": 0, 00:23:40.157 "w_mbytes_per_sec": 0 00:23:40.157 }, 00:23:40.157 "claimed": false, 00:23:40.157 "zoned": false, 00:23:40.157 "supported_io_types": { 00:23:40.157 "read": true, 00:23:40.157 "write": true, 00:23:40.157 "unmap": true, 00:23:40.157 "flush": true, 00:23:40.157 "reset": true, 00:23:40.157 "nvme_admin": false, 00:23:40.157 "nvme_io": false, 00:23:40.157 "nvme_io_md": false, 00:23:40.157 "write_zeroes": true, 00:23:40.157 "zcopy": true, 00:23:40.157 "get_zone_info": false, 00:23:40.157 "zone_management": false, 00:23:40.157 "zone_append": false, 00:23:40.157 "compare": false, 00:23:40.157 "compare_and_write": false, 00:23:40.157 "abort": true, 00:23:40.157 "seek_hole": false, 00:23:40.157 "seek_data": false, 00:23:40.157 "copy": true, 00:23:40.157 "nvme_iov_md": false 00:23:40.157 }, 00:23:40.157 "memory_domains": [ 00:23:40.157 { 00:23:40.157 "dma_device_id": "system", 00:23:40.157 "dma_device_type": 1 00:23:40.157 }, 00:23:40.157 { 00:23:40.157 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:40.157 "dma_device_type": 2 00:23:40.157 } 00:23:40.157 ], 00:23:40.157 "driver_specific": {} 00:23:40.157 } 00:23:40.157 ] 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:40.157 BaseBdev4 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:40.157 [ 00:23:40.157 { 00:23:40.157 "name": "BaseBdev4", 00:23:40.157 "aliases": [ 00:23:40.157 "62a3e67e-d6a4-44d3-b2a1-13739795865f" 00:23:40.157 ], 00:23:40.157 "product_name": "Malloc disk", 00:23:40.157 "block_size": 512, 00:23:40.157 "num_blocks": 65536, 00:23:40.157 "uuid": "62a3e67e-d6a4-44d3-b2a1-13739795865f", 00:23:40.157 "assigned_rate_limits": { 00:23:40.157 "rw_ios_per_sec": 0, 00:23:40.157 "rw_mbytes_per_sec": 0, 00:23:40.157 "r_mbytes_per_sec": 0, 00:23:40.157 "w_mbytes_per_sec": 0 00:23:40.157 }, 00:23:40.157 "claimed": false, 00:23:40.157 "zoned": false, 00:23:40.157 "supported_io_types": { 00:23:40.157 "read": true, 00:23:40.157 "write": true, 00:23:40.157 "unmap": true, 00:23:40.157 "flush": true, 00:23:40.157 "reset": true, 00:23:40.157 "nvme_admin": false, 00:23:40.157 "nvme_io": false, 00:23:40.157 "nvme_io_md": false, 00:23:40.157 "write_zeroes": true, 00:23:40.157 "zcopy": true, 00:23:40.157 "get_zone_info": false, 00:23:40.157 "zone_management": false, 00:23:40.157 "zone_append": false, 00:23:40.157 "compare": false, 00:23:40.157 "compare_and_write": false, 00:23:40.157 "abort": true, 00:23:40.157 "seek_hole": false, 00:23:40.157 "seek_data": false, 00:23:40.157 "copy": true, 00:23:40.157 "nvme_iov_md": false 00:23:40.157 }, 00:23:40.157 "memory_domains": [ 00:23:40.157 { 00:23:40.157 "dma_device_id": "system", 00:23:40.157 "dma_device_type": 1 00:23:40.157 }, 00:23:40.157 { 00:23:40.157 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:40.157 "dma_device_type": 2 00:23:40.157 } 00:23:40.157 ], 00:23:40.157 "driver_specific": {} 00:23:40.157 } 00:23:40.157 ] 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:40.157 [2024-10-01 20:22:35.338863] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:23:40.157 [2024-10-01 20:22:35.339056] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:23:40.157 [2024-10-01 20:22:35.339203] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:23:40.157 [2024-10-01 20:22:35.341815] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:23:40.157 [2024-10-01 20:22:35.341885] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:40.157 "name": "Existed_Raid", 00:23:40.157 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:40.157 "strip_size_kb": 64, 00:23:40.157 "state": "configuring", 00:23:40.157 "raid_level": "raid0", 00:23:40.157 "superblock": false, 00:23:40.157 "num_base_bdevs": 4, 00:23:40.157 "num_base_bdevs_discovered": 3, 00:23:40.157 "num_base_bdevs_operational": 4, 00:23:40.157 "base_bdevs_list": [ 00:23:40.157 { 00:23:40.157 "name": "BaseBdev1", 00:23:40.157 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:40.157 "is_configured": false, 00:23:40.157 "data_offset": 0, 00:23:40.157 "data_size": 0 00:23:40.157 }, 00:23:40.157 { 00:23:40.157 "name": "BaseBdev2", 00:23:40.157 "uuid": "d4af6e2c-79ce-417d-8b5f-476b5d3b4690", 00:23:40.157 "is_configured": true, 00:23:40.157 "data_offset": 0, 00:23:40.157 "data_size": 65536 00:23:40.157 }, 00:23:40.157 { 00:23:40.157 "name": "BaseBdev3", 00:23:40.157 "uuid": "be32d2e4-fe7e-4908-b551-c3301540ed03", 00:23:40.157 "is_configured": true, 00:23:40.157 "data_offset": 0, 00:23:40.157 "data_size": 65536 00:23:40.157 }, 00:23:40.157 { 00:23:40.157 "name": "BaseBdev4", 00:23:40.157 "uuid": "62a3e67e-d6a4-44d3-b2a1-13739795865f", 00:23:40.157 "is_configured": true, 00:23:40.157 "data_offset": 0, 00:23:40.157 "data_size": 65536 00:23:40.157 } 00:23:40.157 ] 00:23:40.157 }' 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:40.157 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:40.724 [2024-10-01 20:22:35.855068] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:40.724 "name": "Existed_Raid", 00:23:40.724 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:40.724 "strip_size_kb": 64, 00:23:40.724 "state": "configuring", 00:23:40.724 "raid_level": "raid0", 00:23:40.724 "superblock": false, 00:23:40.724 "num_base_bdevs": 4, 00:23:40.724 "num_base_bdevs_discovered": 2, 00:23:40.724 "num_base_bdevs_operational": 4, 00:23:40.724 "base_bdevs_list": [ 00:23:40.724 { 00:23:40.724 "name": "BaseBdev1", 00:23:40.724 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:40.724 "is_configured": false, 00:23:40.724 "data_offset": 0, 00:23:40.724 "data_size": 0 00:23:40.724 }, 00:23:40.724 { 00:23:40.724 "name": null, 00:23:40.724 "uuid": "d4af6e2c-79ce-417d-8b5f-476b5d3b4690", 00:23:40.724 "is_configured": false, 00:23:40.724 "data_offset": 0, 00:23:40.724 "data_size": 65536 00:23:40.724 }, 00:23:40.724 { 00:23:40.724 "name": "BaseBdev3", 00:23:40.724 "uuid": "be32d2e4-fe7e-4908-b551-c3301540ed03", 00:23:40.724 "is_configured": true, 00:23:40.724 "data_offset": 0, 00:23:40.724 "data_size": 65536 00:23:40.724 }, 00:23:40.724 { 00:23:40.724 "name": "BaseBdev4", 00:23:40.724 "uuid": "62a3e67e-d6a4-44d3-b2a1-13739795865f", 00:23:40.724 "is_configured": true, 00:23:40.724 "data_offset": 0, 00:23:40.724 "data_size": 65536 00:23:40.724 } 00:23:40.724 ] 00:23:40.724 }' 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:40.724 20:22:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:41.290 [2024-10-01 20:22:36.464295] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:23:41.290 BaseBdev1 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:41.290 [ 00:23:41.290 { 00:23:41.290 "name": "BaseBdev1", 00:23:41.290 "aliases": [ 00:23:41.290 "ba254384-025c-4d9a-86cb-a51c829b5803" 00:23:41.290 ], 00:23:41.290 "product_name": "Malloc disk", 00:23:41.290 "block_size": 512, 00:23:41.290 "num_blocks": 65536, 00:23:41.290 "uuid": "ba254384-025c-4d9a-86cb-a51c829b5803", 00:23:41.290 "assigned_rate_limits": { 00:23:41.290 "rw_ios_per_sec": 0, 00:23:41.290 "rw_mbytes_per_sec": 0, 00:23:41.290 "r_mbytes_per_sec": 0, 00:23:41.290 "w_mbytes_per_sec": 0 00:23:41.290 }, 00:23:41.290 "claimed": true, 00:23:41.290 "claim_type": "exclusive_write", 00:23:41.290 "zoned": false, 00:23:41.290 "supported_io_types": { 00:23:41.290 "read": true, 00:23:41.290 "write": true, 00:23:41.290 "unmap": true, 00:23:41.290 "flush": true, 00:23:41.290 "reset": true, 00:23:41.290 "nvme_admin": false, 00:23:41.290 "nvme_io": false, 00:23:41.290 "nvme_io_md": false, 00:23:41.290 "write_zeroes": true, 00:23:41.290 "zcopy": true, 00:23:41.290 "get_zone_info": false, 00:23:41.290 "zone_management": false, 00:23:41.290 "zone_append": false, 00:23:41.290 "compare": false, 00:23:41.290 "compare_and_write": false, 00:23:41.290 "abort": true, 00:23:41.290 "seek_hole": false, 00:23:41.290 "seek_data": false, 00:23:41.290 "copy": true, 00:23:41.290 "nvme_iov_md": false 00:23:41.290 }, 00:23:41.290 "memory_domains": [ 00:23:41.290 { 00:23:41.290 "dma_device_id": "system", 00:23:41.290 "dma_device_type": 1 00:23:41.290 }, 00:23:41.290 { 00:23:41.290 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:41.290 "dma_device_type": 2 00:23:41.290 } 00:23:41.290 ], 00:23:41.290 "driver_specific": {} 00:23:41.290 } 00:23:41.290 ] 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:41.290 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:41.548 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:41.548 "name": "Existed_Raid", 00:23:41.548 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:41.548 "strip_size_kb": 64, 00:23:41.548 "state": "configuring", 00:23:41.548 "raid_level": "raid0", 00:23:41.548 "superblock": false, 00:23:41.548 "num_base_bdevs": 4, 00:23:41.548 "num_base_bdevs_discovered": 3, 00:23:41.548 "num_base_bdevs_operational": 4, 00:23:41.548 "base_bdevs_list": [ 00:23:41.548 { 00:23:41.548 "name": "BaseBdev1", 00:23:41.548 "uuid": "ba254384-025c-4d9a-86cb-a51c829b5803", 00:23:41.548 "is_configured": true, 00:23:41.548 "data_offset": 0, 00:23:41.548 "data_size": 65536 00:23:41.548 }, 00:23:41.548 { 00:23:41.548 "name": null, 00:23:41.548 "uuid": "d4af6e2c-79ce-417d-8b5f-476b5d3b4690", 00:23:41.548 "is_configured": false, 00:23:41.548 "data_offset": 0, 00:23:41.548 "data_size": 65536 00:23:41.548 }, 00:23:41.548 { 00:23:41.548 "name": "BaseBdev3", 00:23:41.548 "uuid": "be32d2e4-fe7e-4908-b551-c3301540ed03", 00:23:41.548 "is_configured": true, 00:23:41.548 "data_offset": 0, 00:23:41.548 "data_size": 65536 00:23:41.548 }, 00:23:41.548 { 00:23:41.548 "name": "BaseBdev4", 00:23:41.548 "uuid": "62a3e67e-d6a4-44d3-b2a1-13739795865f", 00:23:41.548 "is_configured": true, 00:23:41.548 "data_offset": 0, 00:23:41.548 "data_size": 65536 00:23:41.548 } 00:23:41.548 ] 00:23:41.548 }' 00:23:41.548 20:22:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:41.548 20:22:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:41.805 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:41.805 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:41.805 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:41.805 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:23:41.805 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:42.136 [2024-10-01 20:22:37.092577] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:42.136 "name": "Existed_Raid", 00:23:42.136 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:42.136 "strip_size_kb": 64, 00:23:42.136 "state": "configuring", 00:23:42.136 "raid_level": "raid0", 00:23:42.136 "superblock": false, 00:23:42.136 "num_base_bdevs": 4, 00:23:42.136 "num_base_bdevs_discovered": 2, 00:23:42.136 "num_base_bdevs_operational": 4, 00:23:42.136 "base_bdevs_list": [ 00:23:42.136 { 00:23:42.136 "name": "BaseBdev1", 00:23:42.136 "uuid": "ba254384-025c-4d9a-86cb-a51c829b5803", 00:23:42.136 "is_configured": true, 00:23:42.136 "data_offset": 0, 00:23:42.136 "data_size": 65536 00:23:42.136 }, 00:23:42.136 { 00:23:42.136 "name": null, 00:23:42.136 "uuid": "d4af6e2c-79ce-417d-8b5f-476b5d3b4690", 00:23:42.136 "is_configured": false, 00:23:42.136 "data_offset": 0, 00:23:42.136 "data_size": 65536 00:23:42.136 }, 00:23:42.136 { 00:23:42.136 "name": null, 00:23:42.136 "uuid": "be32d2e4-fe7e-4908-b551-c3301540ed03", 00:23:42.136 "is_configured": false, 00:23:42.136 "data_offset": 0, 00:23:42.136 "data_size": 65536 00:23:42.136 }, 00:23:42.136 { 00:23:42.136 "name": "BaseBdev4", 00:23:42.136 "uuid": "62a3e67e-d6a4-44d3-b2a1-13739795865f", 00:23:42.136 "is_configured": true, 00:23:42.136 "data_offset": 0, 00:23:42.136 "data_size": 65536 00:23:42.136 } 00:23:42.136 ] 00:23:42.136 }' 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:42.136 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:42.394 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:42.394 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:23:42.394 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:42.394 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:42.394 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:42.394 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:23:42.394 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:23:42.394 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:42.394 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:42.394 [2024-10-01 20:22:37.644676] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:42.652 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:42.652 "name": "Existed_Raid", 00:23:42.652 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:42.652 "strip_size_kb": 64, 00:23:42.652 "state": "configuring", 00:23:42.652 "raid_level": "raid0", 00:23:42.652 "superblock": false, 00:23:42.652 "num_base_bdevs": 4, 00:23:42.652 "num_base_bdevs_discovered": 3, 00:23:42.652 "num_base_bdevs_operational": 4, 00:23:42.652 "base_bdevs_list": [ 00:23:42.652 { 00:23:42.652 "name": "BaseBdev1", 00:23:42.652 "uuid": "ba254384-025c-4d9a-86cb-a51c829b5803", 00:23:42.652 "is_configured": true, 00:23:42.652 "data_offset": 0, 00:23:42.652 "data_size": 65536 00:23:42.652 }, 00:23:42.652 { 00:23:42.652 "name": null, 00:23:42.652 "uuid": "d4af6e2c-79ce-417d-8b5f-476b5d3b4690", 00:23:42.652 "is_configured": false, 00:23:42.652 "data_offset": 0, 00:23:42.652 "data_size": 65536 00:23:42.652 }, 00:23:42.652 { 00:23:42.652 "name": "BaseBdev3", 00:23:42.652 "uuid": "be32d2e4-fe7e-4908-b551-c3301540ed03", 00:23:42.652 "is_configured": true, 00:23:42.652 "data_offset": 0, 00:23:42.652 "data_size": 65536 00:23:42.652 }, 00:23:42.652 { 00:23:42.653 "name": "BaseBdev4", 00:23:42.653 "uuid": "62a3e67e-d6a4-44d3-b2a1-13739795865f", 00:23:42.653 "is_configured": true, 00:23:42.653 "data_offset": 0, 00:23:42.653 "data_size": 65536 00:23:42.653 } 00:23:42.653 ] 00:23:42.653 }' 00:23:42.653 20:22:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:42.653 20:22:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:43.220 [2024-10-01 20:22:38.220964] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:43.220 "name": "Existed_Raid", 00:23:43.220 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:43.220 "strip_size_kb": 64, 00:23:43.220 "state": "configuring", 00:23:43.220 "raid_level": "raid0", 00:23:43.220 "superblock": false, 00:23:43.220 "num_base_bdevs": 4, 00:23:43.220 "num_base_bdevs_discovered": 2, 00:23:43.220 "num_base_bdevs_operational": 4, 00:23:43.220 "base_bdevs_list": [ 00:23:43.220 { 00:23:43.220 "name": null, 00:23:43.220 "uuid": "ba254384-025c-4d9a-86cb-a51c829b5803", 00:23:43.220 "is_configured": false, 00:23:43.220 "data_offset": 0, 00:23:43.220 "data_size": 65536 00:23:43.220 }, 00:23:43.220 { 00:23:43.220 "name": null, 00:23:43.220 "uuid": "d4af6e2c-79ce-417d-8b5f-476b5d3b4690", 00:23:43.220 "is_configured": false, 00:23:43.220 "data_offset": 0, 00:23:43.220 "data_size": 65536 00:23:43.220 }, 00:23:43.220 { 00:23:43.220 "name": "BaseBdev3", 00:23:43.220 "uuid": "be32d2e4-fe7e-4908-b551-c3301540ed03", 00:23:43.220 "is_configured": true, 00:23:43.220 "data_offset": 0, 00:23:43.220 "data_size": 65536 00:23:43.220 }, 00:23:43.220 { 00:23:43.220 "name": "BaseBdev4", 00:23:43.220 "uuid": "62a3e67e-d6a4-44d3-b2a1-13739795865f", 00:23:43.220 "is_configured": true, 00:23:43.220 "data_offset": 0, 00:23:43.220 "data_size": 65536 00:23:43.220 } 00:23:43.220 ] 00:23:43.220 }' 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:43.220 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:43.788 [2024-10-01 20:22:38.879252] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:43.788 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:43.788 "name": "Existed_Raid", 00:23:43.788 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:43.788 "strip_size_kb": 64, 00:23:43.788 "state": "configuring", 00:23:43.788 "raid_level": "raid0", 00:23:43.788 "superblock": false, 00:23:43.788 "num_base_bdevs": 4, 00:23:43.788 "num_base_bdevs_discovered": 3, 00:23:43.788 "num_base_bdevs_operational": 4, 00:23:43.788 "base_bdevs_list": [ 00:23:43.788 { 00:23:43.788 "name": null, 00:23:43.788 "uuid": "ba254384-025c-4d9a-86cb-a51c829b5803", 00:23:43.788 "is_configured": false, 00:23:43.788 "data_offset": 0, 00:23:43.788 "data_size": 65536 00:23:43.788 }, 00:23:43.788 { 00:23:43.788 "name": "BaseBdev2", 00:23:43.788 "uuid": "d4af6e2c-79ce-417d-8b5f-476b5d3b4690", 00:23:43.788 "is_configured": true, 00:23:43.788 "data_offset": 0, 00:23:43.788 "data_size": 65536 00:23:43.788 }, 00:23:43.788 { 00:23:43.788 "name": "BaseBdev3", 00:23:43.788 "uuid": "be32d2e4-fe7e-4908-b551-c3301540ed03", 00:23:43.788 "is_configured": true, 00:23:43.788 "data_offset": 0, 00:23:43.788 "data_size": 65536 00:23:43.788 }, 00:23:43.788 { 00:23:43.788 "name": "BaseBdev4", 00:23:43.789 "uuid": "62a3e67e-d6a4-44d3-b2a1-13739795865f", 00:23:43.789 "is_configured": true, 00:23:43.789 "data_offset": 0, 00:23:43.789 "data_size": 65536 00:23:43.789 } 00:23:43.789 ] 00:23:43.789 }' 00:23:43.789 20:22:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:43.789 20:22:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u ba254384-025c-4d9a-86cb-a51c829b5803 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:44.355 [2024-10-01 20:22:39.550691] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:23:44.355 [2024-10-01 20:22:39.550759] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:23:44.355 [2024-10-01 20:22:39.550772] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:23:44.355 [2024-10-01 20:22:39.551380] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:23:44.355 [2024-10-01 20:22:39.551579] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:23:44.355 [2024-10-01 20:22:39.551601] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:23:44.355 [2024-10-01 20:22:39.551910] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:44.355 NewBaseBdev 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:44.355 [ 00:23:44.355 { 00:23:44.355 "name": "NewBaseBdev", 00:23:44.355 "aliases": [ 00:23:44.355 "ba254384-025c-4d9a-86cb-a51c829b5803" 00:23:44.355 ], 00:23:44.355 "product_name": "Malloc disk", 00:23:44.355 "block_size": 512, 00:23:44.355 "num_blocks": 65536, 00:23:44.355 "uuid": "ba254384-025c-4d9a-86cb-a51c829b5803", 00:23:44.355 "assigned_rate_limits": { 00:23:44.355 "rw_ios_per_sec": 0, 00:23:44.355 "rw_mbytes_per_sec": 0, 00:23:44.355 "r_mbytes_per_sec": 0, 00:23:44.355 "w_mbytes_per_sec": 0 00:23:44.355 }, 00:23:44.355 "claimed": true, 00:23:44.355 "claim_type": "exclusive_write", 00:23:44.355 "zoned": false, 00:23:44.355 "supported_io_types": { 00:23:44.355 "read": true, 00:23:44.355 "write": true, 00:23:44.355 "unmap": true, 00:23:44.355 "flush": true, 00:23:44.355 "reset": true, 00:23:44.355 "nvme_admin": false, 00:23:44.355 "nvme_io": false, 00:23:44.355 "nvme_io_md": false, 00:23:44.355 "write_zeroes": true, 00:23:44.355 "zcopy": true, 00:23:44.355 "get_zone_info": false, 00:23:44.355 "zone_management": false, 00:23:44.355 "zone_append": false, 00:23:44.355 "compare": false, 00:23:44.355 "compare_and_write": false, 00:23:44.355 "abort": true, 00:23:44.355 "seek_hole": false, 00:23:44.355 "seek_data": false, 00:23:44.355 "copy": true, 00:23:44.355 "nvme_iov_md": false 00:23:44.355 }, 00:23:44.355 "memory_domains": [ 00:23:44.355 { 00:23:44.355 "dma_device_id": "system", 00:23:44.355 "dma_device_type": 1 00:23:44.355 }, 00:23:44.355 { 00:23:44.355 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:44.355 "dma_device_type": 2 00:23:44.355 } 00:23:44.355 ], 00:23:44.355 "driver_specific": {} 00:23:44.355 } 00:23:44.355 ] 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:44.355 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:44.356 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:44.356 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:44.356 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:44.356 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:44.356 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:44.616 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:44.616 "name": "Existed_Raid", 00:23:44.616 "uuid": "4603b43a-d086-440c-976c-1f7df0a4ba3c", 00:23:44.616 "strip_size_kb": 64, 00:23:44.616 "state": "online", 00:23:44.616 "raid_level": "raid0", 00:23:44.616 "superblock": false, 00:23:44.616 "num_base_bdevs": 4, 00:23:44.616 "num_base_bdevs_discovered": 4, 00:23:44.616 "num_base_bdevs_operational": 4, 00:23:44.616 "base_bdevs_list": [ 00:23:44.616 { 00:23:44.616 "name": "NewBaseBdev", 00:23:44.616 "uuid": "ba254384-025c-4d9a-86cb-a51c829b5803", 00:23:44.616 "is_configured": true, 00:23:44.616 "data_offset": 0, 00:23:44.616 "data_size": 65536 00:23:44.616 }, 00:23:44.616 { 00:23:44.616 "name": "BaseBdev2", 00:23:44.616 "uuid": "d4af6e2c-79ce-417d-8b5f-476b5d3b4690", 00:23:44.616 "is_configured": true, 00:23:44.616 "data_offset": 0, 00:23:44.616 "data_size": 65536 00:23:44.616 }, 00:23:44.616 { 00:23:44.616 "name": "BaseBdev3", 00:23:44.616 "uuid": "be32d2e4-fe7e-4908-b551-c3301540ed03", 00:23:44.616 "is_configured": true, 00:23:44.616 "data_offset": 0, 00:23:44.616 "data_size": 65536 00:23:44.616 }, 00:23:44.616 { 00:23:44.616 "name": "BaseBdev4", 00:23:44.616 "uuid": "62a3e67e-d6a4-44d3-b2a1-13739795865f", 00:23:44.616 "is_configured": true, 00:23:44.616 "data_offset": 0, 00:23:44.616 "data_size": 65536 00:23:44.616 } 00:23:44.616 ] 00:23:44.616 }' 00:23:44.616 20:22:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:44.616 20:22:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:44.874 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:23:44.874 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:23:44.874 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:23:44.874 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:23:44.874 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:23:44.874 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:23:44.874 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:23:44.874 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:23:44.874 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:44.874 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:44.874 [2024-10-01 20:22:40.103378] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:23:45.133 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:45.133 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:45.133 "name": "Existed_Raid", 00:23:45.133 "aliases": [ 00:23:45.133 "4603b43a-d086-440c-976c-1f7df0a4ba3c" 00:23:45.133 ], 00:23:45.133 "product_name": "Raid Volume", 00:23:45.133 "block_size": 512, 00:23:45.133 "num_blocks": 262144, 00:23:45.133 "uuid": "4603b43a-d086-440c-976c-1f7df0a4ba3c", 00:23:45.133 "assigned_rate_limits": { 00:23:45.133 "rw_ios_per_sec": 0, 00:23:45.133 "rw_mbytes_per_sec": 0, 00:23:45.133 "r_mbytes_per_sec": 0, 00:23:45.133 "w_mbytes_per_sec": 0 00:23:45.133 }, 00:23:45.133 "claimed": false, 00:23:45.133 "zoned": false, 00:23:45.133 "supported_io_types": { 00:23:45.133 "read": true, 00:23:45.133 "write": true, 00:23:45.133 "unmap": true, 00:23:45.133 "flush": true, 00:23:45.133 "reset": true, 00:23:45.133 "nvme_admin": false, 00:23:45.133 "nvme_io": false, 00:23:45.133 "nvme_io_md": false, 00:23:45.133 "write_zeroes": true, 00:23:45.133 "zcopy": false, 00:23:45.133 "get_zone_info": false, 00:23:45.133 "zone_management": false, 00:23:45.133 "zone_append": false, 00:23:45.133 "compare": false, 00:23:45.133 "compare_and_write": false, 00:23:45.133 "abort": false, 00:23:45.133 "seek_hole": false, 00:23:45.133 "seek_data": false, 00:23:45.133 "copy": false, 00:23:45.133 "nvme_iov_md": false 00:23:45.133 }, 00:23:45.133 "memory_domains": [ 00:23:45.133 { 00:23:45.133 "dma_device_id": "system", 00:23:45.133 "dma_device_type": 1 00:23:45.133 }, 00:23:45.133 { 00:23:45.133 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:45.133 "dma_device_type": 2 00:23:45.133 }, 00:23:45.133 { 00:23:45.133 "dma_device_id": "system", 00:23:45.133 "dma_device_type": 1 00:23:45.133 }, 00:23:45.133 { 00:23:45.133 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:45.133 "dma_device_type": 2 00:23:45.133 }, 00:23:45.133 { 00:23:45.133 "dma_device_id": "system", 00:23:45.133 "dma_device_type": 1 00:23:45.133 }, 00:23:45.133 { 00:23:45.133 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:45.133 "dma_device_type": 2 00:23:45.133 }, 00:23:45.133 { 00:23:45.133 "dma_device_id": "system", 00:23:45.133 "dma_device_type": 1 00:23:45.133 }, 00:23:45.133 { 00:23:45.133 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:45.133 "dma_device_type": 2 00:23:45.133 } 00:23:45.133 ], 00:23:45.133 "driver_specific": { 00:23:45.133 "raid": { 00:23:45.133 "uuid": "4603b43a-d086-440c-976c-1f7df0a4ba3c", 00:23:45.133 "strip_size_kb": 64, 00:23:45.133 "state": "online", 00:23:45.133 "raid_level": "raid0", 00:23:45.133 "superblock": false, 00:23:45.133 "num_base_bdevs": 4, 00:23:45.133 "num_base_bdevs_discovered": 4, 00:23:45.133 "num_base_bdevs_operational": 4, 00:23:45.133 "base_bdevs_list": [ 00:23:45.133 { 00:23:45.133 "name": "NewBaseBdev", 00:23:45.133 "uuid": "ba254384-025c-4d9a-86cb-a51c829b5803", 00:23:45.133 "is_configured": true, 00:23:45.133 "data_offset": 0, 00:23:45.133 "data_size": 65536 00:23:45.133 }, 00:23:45.133 { 00:23:45.133 "name": "BaseBdev2", 00:23:45.133 "uuid": "d4af6e2c-79ce-417d-8b5f-476b5d3b4690", 00:23:45.133 "is_configured": true, 00:23:45.133 "data_offset": 0, 00:23:45.133 "data_size": 65536 00:23:45.133 }, 00:23:45.133 { 00:23:45.133 "name": "BaseBdev3", 00:23:45.133 "uuid": "be32d2e4-fe7e-4908-b551-c3301540ed03", 00:23:45.133 "is_configured": true, 00:23:45.133 "data_offset": 0, 00:23:45.133 "data_size": 65536 00:23:45.133 }, 00:23:45.133 { 00:23:45.133 "name": "BaseBdev4", 00:23:45.133 "uuid": "62a3e67e-d6a4-44d3-b2a1-13739795865f", 00:23:45.133 "is_configured": true, 00:23:45.133 "data_offset": 0, 00:23:45.133 "data_size": 65536 00:23:45.133 } 00:23:45.133 ] 00:23:45.133 } 00:23:45.133 } 00:23:45.133 }' 00:23:45.133 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:23:45.133 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:23:45.133 BaseBdev2 00:23:45.133 BaseBdev3 00:23:45.133 BaseBdev4' 00:23:45.133 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:45.133 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:23:45.133 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:45.133 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:23:45.133 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:45.134 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:45.393 [2024-10-01 20:22:40.474998] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:23:45.393 [2024-10-01 20:22:40.475036] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:23:45.393 [2024-10-01 20:22:40.475170] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:23:45.393 [2024-10-01 20:22:40.475268] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:23:45.393 [2024-10-01 20:22:40.475284] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 70095 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 70095 ']' 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 70095 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70095 00:23:45.393 killing process with pid 70095 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70095' 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 70095 00:23:45.393 [2024-10-01 20:22:40.515011] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:23:45.393 20:22:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 70095 00:23:45.652 [2024-10-01 20:22:40.850896] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:23:47.561 00:23:47.561 real 0m13.500s 00:23:47.561 user 0m21.719s 00:23:47.561 sys 0m1.973s 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:23:47.561 ************************************ 00:23:47.561 END TEST raid_state_function_test 00:23:47.561 ************************************ 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:23:47.561 20:22:42 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 4 true 00:23:47.561 20:22:42 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:23:47.561 20:22:42 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:23:47.561 20:22:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:23:47.561 ************************************ 00:23:47.561 START TEST raid_state_function_test_sb 00:23:47.561 ************************************ 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 4 true 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:23:47.561 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=70783 00:23:47.562 Process raid pid: 70783 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 70783' 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 70783 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 70783 ']' 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:47.562 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:47.562 20:22:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:47.562 [2024-10-01 20:22:42.770174] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:23:47.562 [2024-10-01 20:22:42.770386] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:47.820 [2024-10-01 20:22:42.935652] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:48.078 [2024-10-01 20:22:43.175316] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:23:48.336 [2024-10-01 20:22:43.367983] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:23:48.336 [2024-10-01 20:22:43.368053] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:23:48.594 20:22:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:48.595 [2024-10-01 20:22:43.796956] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:23:48.595 [2024-10-01 20:22:43.797046] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:23:48.595 [2024-10-01 20:22:43.797065] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:23:48.595 [2024-10-01 20:22:43.797083] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:23:48.595 [2024-10-01 20:22:43.797093] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:23:48.595 [2024-10-01 20:22:43.797111] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:23:48.595 [2024-10-01 20:22:43.797122] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:23:48.595 [2024-10-01 20:22:43.797137] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:48.595 20:22:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:48.852 20:22:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:48.852 "name": "Existed_Raid", 00:23:48.852 "uuid": "308ecb47-73a6-4295-9f7a-f12c294c4055", 00:23:48.852 "strip_size_kb": 64, 00:23:48.852 "state": "configuring", 00:23:48.852 "raid_level": "raid0", 00:23:48.852 "superblock": true, 00:23:48.852 "num_base_bdevs": 4, 00:23:48.852 "num_base_bdevs_discovered": 0, 00:23:48.852 "num_base_bdevs_operational": 4, 00:23:48.852 "base_bdevs_list": [ 00:23:48.852 { 00:23:48.852 "name": "BaseBdev1", 00:23:48.852 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:48.852 "is_configured": false, 00:23:48.852 "data_offset": 0, 00:23:48.852 "data_size": 0 00:23:48.852 }, 00:23:48.852 { 00:23:48.852 "name": "BaseBdev2", 00:23:48.852 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:48.852 "is_configured": false, 00:23:48.852 "data_offset": 0, 00:23:48.852 "data_size": 0 00:23:48.852 }, 00:23:48.852 { 00:23:48.852 "name": "BaseBdev3", 00:23:48.852 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:48.852 "is_configured": false, 00:23:48.852 "data_offset": 0, 00:23:48.852 "data_size": 0 00:23:48.852 }, 00:23:48.852 { 00:23:48.852 "name": "BaseBdev4", 00:23:48.852 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:48.852 "is_configured": false, 00:23:48.852 "data_offset": 0, 00:23:48.852 "data_size": 0 00:23:48.852 } 00:23:48.852 ] 00:23:48.852 }' 00:23:48.852 20:22:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:48.852 20:22:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:49.111 [2024-10-01 20:22:44.297050] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:23:49.111 [2024-10-01 20:22:44.297128] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:49.111 [2024-10-01 20:22:44.305017] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:23:49.111 [2024-10-01 20:22:44.305087] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:23:49.111 [2024-10-01 20:22:44.305103] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:23:49.111 [2024-10-01 20:22:44.305119] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:23:49.111 [2024-10-01 20:22:44.305132] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:23:49.111 [2024-10-01 20:22:44.305147] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:23:49.111 [2024-10-01 20:22:44.305157] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:23:49.111 [2024-10-01 20:22:44.305172] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:49.111 [2024-10-01 20:22:44.351183] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:23:49.111 BaseBdev1 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:49.111 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:49.369 [ 00:23:49.369 { 00:23:49.369 "name": "BaseBdev1", 00:23:49.369 "aliases": [ 00:23:49.369 "8405c12a-4c4b-4ffb-bb49-425f86588bbd" 00:23:49.369 ], 00:23:49.369 "product_name": "Malloc disk", 00:23:49.369 "block_size": 512, 00:23:49.369 "num_blocks": 65536, 00:23:49.369 "uuid": "8405c12a-4c4b-4ffb-bb49-425f86588bbd", 00:23:49.369 "assigned_rate_limits": { 00:23:49.369 "rw_ios_per_sec": 0, 00:23:49.369 "rw_mbytes_per_sec": 0, 00:23:49.369 "r_mbytes_per_sec": 0, 00:23:49.369 "w_mbytes_per_sec": 0 00:23:49.369 }, 00:23:49.369 "claimed": true, 00:23:49.369 "claim_type": "exclusive_write", 00:23:49.369 "zoned": false, 00:23:49.369 "supported_io_types": { 00:23:49.369 "read": true, 00:23:49.369 "write": true, 00:23:49.369 "unmap": true, 00:23:49.369 "flush": true, 00:23:49.369 "reset": true, 00:23:49.369 "nvme_admin": false, 00:23:49.369 "nvme_io": false, 00:23:49.369 "nvme_io_md": false, 00:23:49.369 "write_zeroes": true, 00:23:49.369 "zcopy": true, 00:23:49.369 "get_zone_info": false, 00:23:49.369 "zone_management": false, 00:23:49.369 "zone_append": false, 00:23:49.369 "compare": false, 00:23:49.369 "compare_and_write": false, 00:23:49.369 "abort": true, 00:23:49.369 "seek_hole": false, 00:23:49.369 "seek_data": false, 00:23:49.369 "copy": true, 00:23:49.369 "nvme_iov_md": false 00:23:49.369 }, 00:23:49.369 "memory_domains": [ 00:23:49.369 { 00:23:49.369 "dma_device_id": "system", 00:23:49.369 "dma_device_type": 1 00:23:49.369 }, 00:23:49.369 { 00:23:49.369 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:49.369 "dma_device_type": 2 00:23:49.369 } 00:23:49.369 ], 00:23:49.369 "driver_specific": {} 00:23:49.369 } 00:23:49.369 ] 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:49.369 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:49.369 "name": "Existed_Raid", 00:23:49.369 "uuid": "1dac09fd-f795-41b5-99da-7dec5835d54c", 00:23:49.369 "strip_size_kb": 64, 00:23:49.369 "state": "configuring", 00:23:49.369 "raid_level": "raid0", 00:23:49.369 "superblock": true, 00:23:49.370 "num_base_bdevs": 4, 00:23:49.370 "num_base_bdevs_discovered": 1, 00:23:49.370 "num_base_bdevs_operational": 4, 00:23:49.370 "base_bdevs_list": [ 00:23:49.370 { 00:23:49.370 "name": "BaseBdev1", 00:23:49.370 "uuid": "8405c12a-4c4b-4ffb-bb49-425f86588bbd", 00:23:49.370 "is_configured": true, 00:23:49.370 "data_offset": 2048, 00:23:49.370 "data_size": 63488 00:23:49.370 }, 00:23:49.370 { 00:23:49.370 "name": "BaseBdev2", 00:23:49.370 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:49.370 "is_configured": false, 00:23:49.370 "data_offset": 0, 00:23:49.370 "data_size": 0 00:23:49.370 }, 00:23:49.370 { 00:23:49.370 "name": "BaseBdev3", 00:23:49.370 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:49.370 "is_configured": false, 00:23:49.370 "data_offset": 0, 00:23:49.370 "data_size": 0 00:23:49.370 }, 00:23:49.370 { 00:23:49.370 "name": "BaseBdev4", 00:23:49.370 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:49.370 "is_configured": false, 00:23:49.370 "data_offset": 0, 00:23:49.370 "data_size": 0 00:23:49.370 } 00:23:49.370 ] 00:23:49.370 }' 00:23:49.370 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:49.370 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:49.936 [2024-10-01 20:22:44.911416] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:23:49.936 [2024-10-01 20:22:44.911501] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:49.936 [2024-10-01 20:22:44.919478] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:23:49.936 [2024-10-01 20:22:44.922213] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:23:49.936 [2024-10-01 20:22:44.922304] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:23:49.936 [2024-10-01 20:22:44.922328] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:23:49.936 [2024-10-01 20:22:44.922357] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:23:49.936 [2024-10-01 20:22:44.922367] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:23:49.936 [2024-10-01 20:22:44.922380] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:49.936 "name": "Existed_Raid", 00:23:49.936 "uuid": "900a05ac-1071-4015-b819-16b480c5cc19", 00:23:49.936 "strip_size_kb": 64, 00:23:49.936 "state": "configuring", 00:23:49.936 "raid_level": "raid0", 00:23:49.936 "superblock": true, 00:23:49.936 "num_base_bdevs": 4, 00:23:49.936 "num_base_bdevs_discovered": 1, 00:23:49.936 "num_base_bdevs_operational": 4, 00:23:49.936 "base_bdevs_list": [ 00:23:49.936 { 00:23:49.936 "name": "BaseBdev1", 00:23:49.936 "uuid": "8405c12a-4c4b-4ffb-bb49-425f86588bbd", 00:23:49.936 "is_configured": true, 00:23:49.936 "data_offset": 2048, 00:23:49.936 "data_size": 63488 00:23:49.936 }, 00:23:49.936 { 00:23:49.936 "name": "BaseBdev2", 00:23:49.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:49.936 "is_configured": false, 00:23:49.936 "data_offset": 0, 00:23:49.936 "data_size": 0 00:23:49.936 }, 00:23:49.936 { 00:23:49.936 "name": "BaseBdev3", 00:23:49.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:49.936 "is_configured": false, 00:23:49.936 "data_offset": 0, 00:23:49.936 "data_size": 0 00:23:49.936 }, 00:23:49.936 { 00:23:49.936 "name": "BaseBdev4", 00:23:49.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:49.936 "is_configured": false, 00:23:49.936 "data_offset": 0, 00:23:49.936 "data_size": 0 00:23:49.936 } 00:23:49.936 ] 00:23:49.936 }' 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:49.936 20:22:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:50.214 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:23:50.214 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:50.214 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:50.474 [2024-10-01 20:22:45.486555] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:23:50.474 BaseBdev2 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:50.474 [ 00:23:50.474 { 00:23:50.474 "name": "BaseBdev2", 00:23:50.474 "aliases": [ 00:23:50.474 "620008da-ee91-490f-9c00-3a7f2b4a3364" 00:23:50.474 ], 00:23:50.474 "product_name": "Malloc disk", 00:23:50.474 "block_size": 512, 00:23:50.474 "num_blocks": 65536, 00:23:50.474 "uuid": "620008da-ee91-490f-9c00-3a7f2b4a3364", 00:23:50.474 "assigned_rate_limits": { 00:23:50.474 "rw_ios_per_sec": 0, 00:23:50.474 "rw_mbytes_per_sec": 0, 00:23:50.474 "r_mbytes_per_sec": 0, 00:23:50.474 "w_mbytes_per_sec": 0 00:23:50.474 }, 00:23:50.474 "claimed": true, 00:23:50.474 "claim_type": "exclusive_write", 00:23:50.474 "zoned": false, 00:23:50.474 "supported_io_types": { 00:23:50.474 "read": true, 00:23:50.474 "write": true, 00:23:50.474 "unmap": true, 00:23:50.474 "flush": true, 00:23:50.474 "reset": true, 00:23:50.474 "nvme_admin": false, 00:23:50.474 "nvme_io": false, 00:23:50.474 "nvme_io_md": false, 00:23:50.474 "write_zeroes": true, 00:23:50.474 "zcopy": true, 00:23:50.474 "get_zone_info": false, 00:23:50.474 "zone_management": false, 00:23:50.474 "zone_append": false, 00:23:50.474 "compare": false, 00:23:50.474 "compare_and_write": false, 00:23:50.474 "abort": true, 00:23:50.474 "seek_hole": false, 00:23:50.474 "seek_data": false, 00:23:50.474 "copy": true, 00:23:50.474 "nvme_iov_md": false 00:23:50.474 }, 00:23:50.474 "memory_domains": [ 00:23:50.474 { 00:23:50.474 "dma_device_id": "system", 00:23:50.474 "dma_device_type": 1 00:23:50.474 }, 00:23:50.474 { 00:23:50.474 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:50.474 "dma_device_type": 2 00:23:50.474 } 00:23:50.474 ], 00:23:50.474 "driver_specific": {} 00:23:50.474 } 00:23:50.474 ] 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:50.474 "name": "Existed_Raid", 00:23:50.474 "uuid": "900a05ac-1071-4015-b819-16b480c5cc19", 00:23:50.474 "strip_size_kb": 64, 00:23:50.474 "state": "configuring", 00:23:50.474 "raid_level": "raid0", 00:23:50.474 "superblock": true, 00:23:50.474 "num_base_bdevs": 4, 00:23:50.474 "num_base_bdevs_discovered": 2, 00:23:50.474 "num_base_bdevs_operational": 4, 00:23:50.474 "base_bdevs_list": [ 00:23:50.474 { 00:23:50.474 "name": "BaseBdev1", 00:23:50.474 "uuid": "8405c12a-4c4b-4ffb-bb49-425f86588bbd", 00:23:50.474 "is_configured": true, 00:23:50.474 "data_offset": 2048, 00:23:50.474 "data_size": 63488 00:23:50.474 }, 00:23:50.474 { 00:23:50.474 "name": "BaseBdev2", 00:23:50.474 "uuid": "620008da-ee91-490f-9c00-3a7f2b4a3364", 00:23:50.474 "is_configured": true, 00:23:50.474 "data_offset": 2048, 00:23:50.474 "data_size": 63488 00:23:50.474 }, 00:23:50.474 { 00:23:50.474 "name": "BaseBdev3", 00:23:50.474 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:50.474 "is_configured": false, 00:23:50.474 "data_offset": 0, 00:23:50.474 "data_size": 0 00:23:50.474 }, 00:23:50.474 { 00:23:50.474 "name": "BaseBdev4", 00:23:50.474 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:50.474 "is_configured": false, 00:23:50.474 "data_offset": 0, 00:23:50.474 "data_size": 0 00:23:50.474 } 00:23:50.474 ] 00:23:50.474 }' 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:50.474 20:22:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:51.041 [2024-10-01 20:22:46.058713] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:23:51.041 BaseBdev3 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:51.041 [ 00:23:51.041 { 00:23:51.041 "name": "BaseBdev3", 00:23:51.041 "aliases": [ 00:23:51.041 "3d445f10-447c-46d2-a078-06800f9bbb5b" 00:23:51.041 ], 00:23:51.041 "product_name": "Malloc disk", 00:23:51.041 "block_size": 512, 00:23:51.041 "num_blocks": 65536, 00:23:51.041 "uuid": "3d445f10-447c-46d2-a078-06800f9bbb5b", 00:23:51.041 "assigned_rate_limits": { 00:23:51.041 "rw_ios_per_sec": 0, 00:23:51.041 "rw_mbytes_per_sec": 0, 00:23:51.041 "r_mbytes_per_sec": 0, 00:23:51.041 "w_mbytes_per_sec": 0 00:23:51.041 }, 00:23:51.041 "claimed": true, 00:23:51.041 "claim_type": "exclusive_write", 00:23:51.041 "zoned": false, 00:23:51.041 "supported_io_types": { 00:23:51.041 "read": true, 00:23:51.041 "write": true, 00:23:51.041 "unmap": true, 00:23:51.041 "flush": true, 00:23:51.041 "reset": true, 00:23:51.041 "nvme_admin": false, 00:23:51.041 "nvme_io": false, 00:23:51.041 "nvme_io_md": false, 00:23:51.041 "write_zeroes": true, 00:23:51.041 "zcopy": true, 00:23:51.041 "get_zone_info": false, 00:23:51.041 "zone_management": false, 00:23:51.041 "zone_append": false, 00:23:51.041 "compare": false, 00:23:51.041 "compare_and_write": false, 00:23:51.041 "abort": true, 00:23:51.041 "seek_hole": false, 00:23:51.041 "seek_data": false, 00:23:51.041 "copy": true, 00:23:51.041 "nvme_iov_md": false 00:23:51.041 }, 00:23:51.041 "memory_domains": [ 00:23:51.041 { 00:23:51.041 "dma_device_id": "system", 00:23:51.041 "dma_device_type": 1 00:23:51.041 }, 00:23:51.041 { 00:23:51.041 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:51.041 "dma_device_type": 2 00:23:51.041 } 00:23:51.041 ], 00:23:51.041 "driver_specific": {} 00:23:51.041 } 00:23:51.041 ] 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:51.041 "name": "Existed_Raid", 00:23:51.041 "uuid": "900a05ac-1071-4015-b819-16b480c5cc19", 00:23:51.041 "strip_size_kb": 64, 00:23:51.041 "state": "configuring", 00:23:51.041 "raid_level": "raid0", 00:23:51.041 "superblock": true, 00:23:51.041 "num_base_bdevs": 4, 00:23:51.041 "num_base_bdevs_discovered": 3, 00:23:51.041 "num_base_bdevs_operational": 4, 00:23:51.041 "base_bdevs_list": [ 00:23:51.041 { 00:23:51.041 "name": "BaseBdev1", 00:23:51.041 "uuid": "8405c12a-4c4b-4ffb-bb49-425f86588bbd", 00:23:51.041 "is_configured": true, 00:23:51.041 "data_offset": 2048, 00:23:51.041 "data_size": 63488 00:23:51.041 }, 00:23:51.041 { 00:23:51.041 "name": "BaseBdev2", 00:23:51.041 "uuid": "620008da-ee91-490f-9c00-3a7f2b4a3364", 00:23:51.041 "is_configured": true, 00:23:51.041 "data_offset": 2048, 00:23:51.041 "data_size": 63488 00:23:51.041 }, 00:23:51.041 { 00:23:51.041 "name": "BaseBdev3", 00:23:51.041 "uuid": "3d445f10-447c-46d2-a078-06800f9bbb5b", 00:23:51.041 "is_configured": true, 00:23:51.041 "data_offset": 2048, 00:23:51.041 "data_size": 63488 00:23:51.041 }, 00:23:51.041 { 00:23:51.041 "name": "BaseBdev4", 00:23:51.041 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:51.041 "is_configured": false, 00:23:51.041 "data_offset": 0, 00:23:51.041 "data_size": 0 00:23:51.041 } 00:23:51.041 ] 00:23:51.041 }' 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:51.041 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:51.608 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:23:51.608 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:51.608 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:51.608 [2024-10-01 20:22:46.658252] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:23:51.608 [2024-10-01 20:22:46.658627] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:23:51.608 [2024-10-01 20:22:46.658662] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:23:51.608 [2024-10-01 20:22:46.659047] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:23:51.608 BaseBdev4 00:23:51.608 [2024-10-01 20:22:46.659274] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:23:51.608 [2024-10-01 20:22:46.659305] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:23:51.608 [2024-10-01 20:22:46.659482] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:51.608 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:51.608 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:23:51.608 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:23:51.608 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:51.608 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:23:51.608 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:51.608 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:51.609 [ 00:23:51.609 { 00:23:51.609 "name": "BaseBdev4", 00:23:51.609 "aliases": [ 00:23:51.609 "bd177b6d-308f-4a9b-b1f1-cbe17fec707e" 00:23:51.609 ], 00:23:51.609 "product_name": "Malloc disk", 00:23:51.609 "block_size": 512, 00:23:51.609 "num_blocks": 65536, 00:23:51.609 "uuid": "bd177b6d-308f-4a9b-b1f1-cbe17fec707e", 00:23:51.609 "assigned_rate_limits": { 00:23:51.609 "rw_ios_per_sec": 0, 00:23:51.609 "rw_mbytes_per_sec": 0, 00:23:51.609 "r_mbytes_per_sec": 0, 00:23:51.609 "w_mbytes_per_sec": 0 00:23:51.609 }, 00:23:51.609 "claimed": true, 00:23:51.609 "claim_type": "exclusive_write", 00:23:51.609 "zoned": false, 00:23:51.609 "supported_io_types": { 00:23:51.609 "read": true, 00:23:51.609 "write": true, 00:23:51.609 "unmap": true, 00:23:51.609 "flush": true, 00:23:51.609 "reset": true, 00:23:51.609 "nvme_admin": false, 00:23:51.609 "nvme_io": false, 00:23:51.609 "nvme_io_md": false, 00:23:51.609 "write_zeroes": true, 00:23:51.609 "zcopy": true, 00:23:51.609 "get_zone_info": false, 00:23:51.609 "zone_management": false, 00:23:51.609 "zone_append": false, 00:23:51.609 "compare": false, 00:23:51.609 "compare_and_write": false, 00:23:51.609 "abort": true, 00:23:51.609 "seek_hole": false, 00:23:51.609 "seek_data": false, 00:23:51.609 "copy": true, 00:23:51.609 "nvme_iov_md": false 00:23:51.609 }, 00:23:51.609 "memory_domains": [ 00:23:51.609 { 00:23:51.609 "dma_device_id": "system", 00:23:51.609 "dma_device_type": 1 00:23:51.609 }, 00:23:51.609 { 00:23:51.609 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:51.609 "dma_device_type": 2 00:23:51.609 } 00:23:51.609 ], 00:23:51.609 "driver_specific": {} 00:23:51.609 } 00:23:51.609 ] 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:51.609 "name": "Existed_Raid", 00:23:51.609 "uuid": "900a05ac-1071-4015-b819-16b480c5cc19", 00:23:51.609 "strip_size_kb": 64, 00:23:51.609 "state": "online", 00:23:51.609 "raid_level": "raid0", 00:23:51.609 "superblock": true, 00:23:51.609 "num_base_bdevs": 4, 00:23:51.609 "num_base_bdevs_discovered": 4, 00:23:51.609 "num_base_bdevs_operational": 4, 00:23:51.609 "base_bdevs_list": [ 00:23:51.609 { 00:23:51.609 "name": "BaseBdev1", 00:23:51.609 "uuid": "8405c12a-4c4b-4ffb-bb49-425f86588bbd", 00:23:51.609 "is_configured": true, 00:23:51.609 "data_offset": 2048, 00:23:51.609 "data_size": 63488 00:23:51.609 }, 00:23:51.609 { 00:23:51.609 "name": "BaseBdev2", 00:23:51.609 "uuid": "620008da-ee91-490f-9c00-3a7f2b4a3364", 00:23:51.609 "is_configured": true, 00:23:51.609 "data_offset": 2048, 00:23:51.609 "data_size": 63488 00:23:51.609 }, 00:23:51.609 { 00:23:51.609 "name": "BaseBdev3", 00:23:51.609 "uuid": "3d445f10-447c-46d2-a078-06800f9bbb5b", 00:23:51.609 "is_configured": true, 00:23:51.609 "data_offset": 2048, 00:23:51.609 "data_size": 63488 00:23:51.609 }, 00:23:51.609 { 00:23:51.609 "name": "BaseBdev4", 00:23:51.609 "uuid": "bd177b6d-308f-4a9b-b1f1-cbe17fec707e", 00:23:51.609 "is_configured": true, 00:23:51.609 "data_offset": 2048, 00:23:51.609 "data_size": 63488 00:23:51.609 } 00:23:51.609 ] 00:23:51.609 }' 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:51.609 20:22:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:52.175 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:23:52.175 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:23:52.175 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:23:52.175 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:23:52.175 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:23:52.175 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:23:52.175 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:23:52.175 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:52.175 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:23:52.175 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:52.176 [2024-10-01 20:22:47.210920] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:52.176 "name": "Existed_Raid", 00:23:52.176 "aliases": [ 00:23:52.176 "900a05ac-1071-4015-b819-16b480c5cc19" 00:23:52.176 ], 00:23:52.176 "product_name": "Raid Volume", 00:23:52.176 "block_size": 512, 00:23:52.176 "num_blocks": 253952, 00:23:52.176 "uuid": "900a05ac-1071-4015-b819-16b480c5cc19", 00:23:52.176 "assigned_rate_limits": { 00:23:52.176 "rw_ios_per_sec": 0, 00:23:52.176 "rw_mbytes_per_sec": 0, 00:23:52.176 "r_mbytes_per_sec": 0, 00:23:52.176 "w_mbytes_per_sec": 0 00:23:52.176 }, 00:23:52.176 "claimed": false, 00:23:52.176 "zoned": false, 00:23:52.176 "supported_io_types": { 00:23:52.176 "read": true, 00:23:52.176 "write": true, 00:23:52.176 "unmap": true, 00:23:52.176 "flush": true, 00:23:52.176 "reset": true, 00:23:52.176 "nvme_admin": false, 00:23:52.176 "nvme_io": false, 00:23:52.176 "nvme_io_md": false, 00:23:52.176 "write_zeroes": true, 00:23:52.176 "zcopy": false, 00:23:52.176 "get_zone_info": false, 00:23:52.176 "zone_management": false, 00:23:52.176 "zone_append": false, 00:23:52.176 "compare": false, 00:23:52.176 "compare_and_write": false, 00:23:52.176 "abort": false, 00:23:52.176 "seek_hole": false, 00:23:52.176 "seek_data": false, 00:23:52.176 "copy": false, 00:23:52.176 "nvme_iov_md": false 00:23:52.176 }, 00:23:52.176 "memory_domains": [ 00:23:52.176 { 00:23:52.176 "dma_device_id": "system", 00:23:52.176 "dma_device_type": 1 00:23:52.176 }, 00:23:52.176 { 00:23:52.176 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:52.176 "dma_device_type": 2 00:23:52.176 }, 00:23:52.176 { 00:23:52.176 "dma_device_id": "system", 00:23:52.176 "dma_device_type": 1 00:23:52.176 }, 00:23:52.176 { 00:23:52.176 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:52.176 "dma_device_type": 2 00:23:52.176 }, 00:23:52.176 { 00:23:52.176 "dma_device_id": "system", 00:23:52.176 "dma_device_type": 1 00:23:52.176 }, 00:23:52.176 { 00:23:52.176 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:52.176 "dma_device_type": 2 00:23:52.176 }, 00:23:52.176 { 00:23:52.176 "dma_device_id": "system", 00:23:52.176 "dma_device_type": 1 00:23:52.176 }, 00:23:52.176 { 00:23:52.176 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:52.176 "dma_device_type": 2 00:23:52.176 } 00:23:52.176 ], 00:23:52.176 "driver_specific": { 00:23:52.176 "raid": { 00:23:52.176 "uuid": "900a05ac-1071-4015-b819-16b480c5cc19", 00:23:52.176 "strip_size_kb": 64, 00:23:52.176 "state": "online", 00:23:52.176 "raid_level": "raid0", 00:23:52.176 "superblock": true, 00:23:52.176 "num_base_bdevs": 4, 00:23:52.176 "num_base_bdevs_discovered": 4, 00:23:52.176 "num_base_bdevs_operational": 4, 00:23:52.176 "base_bdevs_list": [ 00:23:52.176 { 00:23:52.176 "name": "BaseBdev1", 00:23:52.176 "uuid": "8405c12a-4c4b-4ffb-bb49-425f86588bbd", 00:23:52.176 "is_configured": true, 00:23:52.176 "data_offset": 2048, 00:23:52.176 "data_size": 63488 00:23:52.176 }, 00:23:52.176 { 00:23:52.176 "name": "BaseBdev2", 00:23:52.176 "uuid": "620008da-ee91-490f-9c00-3a7f2b4a3364", 00:23:52.176 "is_configured": true, 00:23:52.176 "data_offset": 2048, 00:23:52.176 "data_size": 63488 00:23:52.176 }, 00:23:52.176 { 00:23:52.176 "name": "BaseBdev3", 00:23:52.176 "uuid": "3d445f10-447c-46d2-a078-06800f9bbb5b", 00:23:52.176 "is_configured": true, 00:23:52.176 "data_offset": 2048, 00:23:52.176 "data_size": 63488 00:23:52.176 }, 00:23:52.176 { 00:23:52.176 "name": "BaseBdev4", 00:23:52.176 "uuid": "bd177b6d-308f-4a9b-b1f1-cbe17fec707e", 00:23:52.176 "is_configured": true, 00:23:52.176 "data_offset": 2048, 00:23:52.176 "data_size": 63488 00:23:52.176 } 00:23:52.176 ] 00:23:52.176 } 00:23:52.176 } 00:23:52.176 }' 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:23:52.176 BaseBdev2 00:23:52.176 BaseBdev3 00:23:52.176 BaseBdev4' 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:52.176 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:52.435 [2024-10-01 20:22:47.586649] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:23:52.435 [2024-10-01 20:22:47.586839] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:23:52.435 [2024-10-01 20:22:47.587016] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:52.435 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:52.694 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:52.694 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:52.694 "name": "Existed_Raid", 00:23:52.694 "uuid": "900a05ac-1071-4015-b819-16b480c5cc19", 00:23:52.694 "strip_size_kb": 64, 00:23:52.694 "state": "offline", 00:23:52.694 "raid_level": "raid0", 00:23:52.694 "superblock": true, 00:23:52.694 "num_base_bdevs": 4, 00:23:52.694 "num_base_bdevs_discovered": 3, 00:23:52.694 "num_base_bdevs_operational": 3, 00:23:52.694 "base_bdevs_list": [ 00:23:52.694 { 00:23:52.694 "name": null, 00:23:52.694 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:52.694 "is_configured": false, 00:23:52.694 "data_offset": 0, 00:23:52.694 "data_size": 63488 00:23:52.694 }, 00:23:52.694 { 00:23:52.694 "name": "BaseBdev2", 00:23:52.694 "uuid": "620008da-ee91-490f-9c00-3a7f2b4a3364", 00:23:52.694 "is_configured": true, 00:23:52.694 "data_offset": 2048, 00:23:52.694 "data_size": 63488 00:23:52.694 }, 00:23:52.694 { 00:23:52.694 "name": "BaseBdev3", 00:23:52.694 "uuid": "3d445f10-447c-46d2-a078-06800f9bbb5b", 00:23:52.694 "is_configured": true, 00:23:52.694 "data_offset": 2048, 00:23:52.694 "data_size": 63488 00:23:52.694 }, 00:23:52.694 { 00:23:52.694 "name": "BaseBdev4", 00:23:52.694 "uuid": "bd177b6d-308f-4a9b-b1f1-cbe17fec707e", 00:23:52.694 "is_configured": true, 00:23:52.694 "data_offset": 2048, 00:23:52.694 "data_size": 63488 00:23:52.694 } 00:23:52.694 ] 00:23:52.694 }' 00:23:52.694 20:22:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:52.694 20:22:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:52.953 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:23:52.953 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:23:52.953 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:23:52.953 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:52.953 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:52.953 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.211 [2024-10-01 20:22:48.246277] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.211 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.211 [2024-10-01 20:22:48.398381] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.470 [2024-10-01 20:22:48.548797] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:23:53.470 [2024-10-01 20:22:48.548860] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.470 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.730 BaseBdev2 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.730 [ 00:23:53.730 { 00:23:53.730 "name": "BaseBdev2", 00:23:53.730 "aliases": [ 00:23:53.730 "8f706d88-5059-42dd-b117-0b16828b2a07" 00:23:53.730 ], 00:23:53.730 "product_name": "Malloc disk", 00:23:53.730 "block_size": 512, 00:23:53.730 "num_blocks": 65536, 00:23:53.730 "uuid": "8f706d88-5059-42dd-b117-0b16828b2a07", 00:23:53.730 "assigned_rate_limits": { 00:23:53.730 "rw_ios_per_sec": 0, 00:23:53.730 "rw_mbytes_per_sec": 0, 00:23:53.730 "r_mbytes_per_sec": 0, 00:23:53.730 "w_mbytes_per_sec": 0 00:23:53.730 }, 00:23:53.730 "claimed": false, 00:23:53.730 "zoned": false, 00:23:53.730 "supported_io_types": { 00:23:53.730 "read": true, 00:23:53.730 "write": true, 00:23:53.730 "unmap": true, 00:23:53.730 "flush": true, 00:23:53.730 "reset": true, 00:23:53.730 "nvme_admin": false, 00:23:53.730 "nvme_io": false, 00:23:53.730 "nvme_io_md": false, 00:23:53.730 "write_zeroes": true, 00:23:53.730 "zcopy": true, 00:23:53.730 "get_zone_info": false, 00:23:53.730 "zone_management": false, 00:23:53.730 "zone_append": false, 00:23:53.730 "compare": false, 00:23:53.730 "compare_and_write": false, 00:23:53.730 "abort": true, 00:23:53.730 "seek_hole": false, 00:23:53.730 "seek_data": false, 00:23:53.730 "copy": true, 00:23:53.730 "nvme_iov_md": false 00:23:53.730 }, 00:23:53.730 "memory_domains": [ 00:23:53.730 { 00:23:53.730 "dma_device_id": "system", 00:23:53.730 "dma_device_type": 1 00:23:53.730 }, 00:23:53.730 { 00:23:53.730 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:53.730 "dma_device_type": 2 00:23:53.730 } 00:23:53.730 ], 00:23:53.730 "driver_specific": {} 00:23:53.730 } 00:23:53.730 ] 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.730 BaseBdev3 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:53.730 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.731 [ 00:23:53.731 { 00:23:53.731 "name": "BaseBdev3", 00:23:53.731 "aliases": [ 00:23:53.731 "bd193d7a-a543-4daf-85df-a34a3b98a26a" 00:23:53.731 ], 00:23:53.731 "product_name": "Malloc disk", 00:23:53.731 "block_size": 512, 00:23:53.731 "num_blocks": 65536, 00:23:53.731 "uuid": "bd193d7a-a543-4daf-85df-a34a3b98a26a", 00:23:53.731 "assigned_rate_limits": { 00:23:53.731 "rw_ios_per_sec": 0, 00:23:53.731 "rw_mbytes_per_sec": 0, 00:23:53.731 "r_mbytes_per_sec": 0, 00:23:53.731 "w_mbytes_per_sec": 0 00:23:53.731 }, 00:23:53.731 "claimed": false, 00:23:53.731 "zoned": false, 00:23:53.731 "supported_io_types": { 00:23:53.731 "read": true, 00:23:53.731 "write": true, 00:23:53.731 "unmap": true, 00:23:53.731 "flush": true, 00:23:53.731 "reset": true, 00:23:53.731 "nvme_admin": false, 00:23:53.731 "nvme_io": false, 00:23:53.731 "nvme_io_md": false, 00:23:53.731 "write_zeroes": true, 00:23:53.731 "zcopy": true, 00:23:53.731 "get_zone_info": false, 00:23:53.731 "zone_management": false, 00:23:53.731 "zone_append": false, 00:23:53.731 "compare": false, 00:23:53.731 "compare_and_write": false, 00:23:53.731 "abort": true, 00:23:53.731 "seek_hole": false, 00:23:53.731 "seek_data": false, 00:23:53.731 "copy": true, 00:23:53.731 "nvme_iov_md": false 00:23:53.731 }, 00:23:53.731 "memory_domains": [ 00:23:53.731 { 00:23:53.731 "dma_device_id": "system", 00:23:53.731 "dma_device_type": 1 00:23:53.731 }, 00:23:53.731 { 00:23:53.731 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:53.731 "dma_device_type": 2 00:23:53.731 } 00:23:53.731 ], 00:23:53.731 "driver_specific": {} 00:23:53.731 } 00:23:53.731 ] 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.731 BaseBdev4 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.731 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.731 [ 00:23:53.731 { 00:23:53.731 "name": "BaseBdev4", 00:23:53.731 "aliases": [ 00:23:53.731 "104e730e-f926-4e5f-80c3-e17e1d838155" 00:23:53.731 ], 00:23:53.731 "product_name": "Malloc disk", 00:23:53.731 "block_size": 512, 00:23:53.731 "num_blocks": 65536, 00:23:53.731 "uuid": "104e730e-f926-4e5f-80c3-e17e1d838155", 00:23:53.731 "assigned_rate_limits": { 00:23:53.731 "rw_ios_per_sec": 0, 00:23:53.731 "rw_mbytes_per_sec": 0, 00:23:53.731 "r_mbytes_per_sec": 0, 00:23:53.731 "w_mbytes_per_sec": 0 00:23:53.731 }, 00:23:53.731 "claimed": false, 00:23:53.731 "zoned": false, 00:23:53.731 "supported_io_types": { 00:23:53.731 "read": true, 00:23:53.731 "write": true, 00:23:53.731 "unmap": true, 00:23:53.731 "flush": true, 00:23:53.731 "reset": true, 00:23:53.731 "nvme_admin": false, 00:23:53.731 "nvme_io": false, 00:23:53.731 "nvme_io_md": false, 00:23:53.731 "write_zeroes": true, 00:23:53.731 "zcopy": true, 00:23:53.731 "get_zone_info": false, 00:23:53.731 "zone_management": false, 00:23:53.731 "zone_append": false, 00:23:53.731 "compare": false, 00:23:53.731 "compare_and_write": false, 00:23:53.731 "abort": true, 00:23:53.731 "seek_hole": false, 00:23:53.732 "seek_data": false, 00:23:53.732 "copy": true, 00:23:53.732 "nvme_iov_md": false 00:23:53.732 }, 00:23:53.732 "memory_domains": [ 00:23:53.732 { 00:23:53.732 "dma_device_id": "system", 00:23:53.732 "dma_device_type": 1 00:23:53.732 }, 00:23:53.732 { 00:23:53.732 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:53.732 "dma_device_type": 2 00:23:53.732 } 00:23:53.732 ], 00:23:53.732 "driver_specific": {} 00:23:53.732 } 00:23:53.732 ] 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.732 [2024-10-01 20:22:48.931794] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:23:53.732 [2024-10-01 20:22:48.931851] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:23:53.732 [2024-10-01 20:22:48.931900] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:23:53.732 [2024-10-01 20:22:48.934535] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:23:53.732 [2024-10-01 20:22:48.934603] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:53.732 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:53.996 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:53.996 "name": "Existed_Raid", 00:23:53.996 "uuid": "8e49c8d0-f3d8-4493-89b5-ea9e4e4c3366", 00:23:53.996 "strip_size_kb": 64, 00:23:53.996 "state": "configuring", 00:23:53.996 "raid_level": "raid0", 00:23:53.996 "superblock": true, 00:23:53.996 "num_base_bdevs": 4, 00:23:53.996 "num_base_bdevs_discovered": 3, 00:23:53.996 "num_base_bdevs_operational": 4, 00:23:53.996 "base_bdevs_list": [ 00:23:53.996 { 00:23:53.996 "name": "BaseBdev1", 00:23:53.996 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:53.996 "is_configured": false, 00:23:53.996 "data_offset": 0, 00:23:53.996 "data_size": 0 00:23:53.996 }, 00:23:53.996 { 00:23:53.996 "name": "BaseBdev2", 00:23:53.996 "uuid": "8f706d88-5059-42dd-b117-0b16828b2a07", 00:23:53.996 "is_configured": true, 00:23:53.996 "data_offset": 2048, 00:23:53.996 "data_size": 63488 00:23:53.996 }, 00:23:53.996 { 00:23:53.996 "name": "BaseBdev3", 00:23:53.996 "uuid": "bd193d7a-a543-4daf-85df-a34a3b98a26a", 00:23:53.996 "is_configured": true, 00:23:53.996 "data_offset": 2048, 00:23:53.996 "data_size": 63488 00:23:53.996 }, 00:23:53.996 { 00:23:53.996 "name": "BaseBdev4", 00:23:53.996 "uuid": "104e730e-f926-4e5f-80c3-e17e1d838155", 00:23:53.996 "is_configured": true, 00:23:53.996 "data_offset": 2048, 00:23:53.996 "data_size": 63488 00:23:53.996 } 00:23:53.996 ] 00:23:53.996 }' 00:23:53.996 20:22:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:53.996 20:22:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:54.262 [2024-10-01 20:22:49.443969] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:54.262 "name": "Existed_Raid", 00:23:54.262 "uuid": "8e49c8d0-f3d8-4493-89b5-ea9e4e4c3366", 00:23:54.262 "strip_size_kb": 64, 00:23:54.262 "state": "configuring", 00:23:54.262 "raid_level": "raid0", 00:23:54.262 "superblock": true, 00:23:54.262 "num_base_bdevs": 4, 00:23:54.262 "num_base_bdevs_discovered": 2, 00:23:54.262 "num_base_bdevs_operational": 4, 00:23:54.262 "base_bdevs_list": [ 00:23:54.262 { 00:23:54.262 "name": "BaseBdev1", 00:23:54.262 "uuid": "00000000-0000-0000-0000-000000000000", 00:23:54.262 "is_configured": false, 00:23:54.262 "data_offset": 0, 00:23:54.262 "data_size": 0 00:23:54.262 }, 00:23:54.262 { 00:23:54.262 "name": null, 00:23:54.262 "uuid": "8f706d88-5059-42dd-b117-0b16828b2a07", 00:23:54.262 "is_configured": false, 00:23:54.262 "data_offset": 0, 00:23:54.262 "data_size": 63488 00:23:54.262 }, 00:23:54.262 { 00:23:54.262 "name": "BaseBdev3", 00:23:54.262 "uuid": "bd193d7a-a543-4daf-85df-a34a3b98a26a", 00:23:54.262 "is_configured": true, 00:23:54.262 "data_offset": 2048, 00:23:54.262 "data_size": 63488 00:23:54.262 }, 00:23:54.262 { 00:23:54.262 "name": "BaseBdev4", 00:23:54.262 "uuid": "104e730e-f926-4e5f-80c3-e17e1d838155", 00:23:54.262 "is_configured": true, 00:23:54.262 "data_offset": 2048, 00:23:54.262 "data_size": 63488 00:23:54.262 } 00:23:54.262 ] 00:23:54.262 }' 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:54.262 20:22:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:54.830 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:23:54.830 20:22:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:54.830 20:22:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:54.830 20:22:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:54.830 20:22:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:54.830 [2024-10-01 20:22:50.047035] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:23:54.830 BaseBdev1 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:54.830 [ 00:23:54.830 { 00:23:54.830 "name": "BaseBdev1", 00:23:54.830 "aliases": [ 00:23:54.830 "5b2f6391-eba3-4830-8ce9-c78fe025995e" 00:23:54.830 ], 00:23:54.830 "product_name": "Malloc disk", 00:23:54.830 "block_size": 512, 00:23:54.830 "num_blocks": 65536, 00:23:54.830 "uuid": "5b2f6391-eba3-4830-8ce9-c78fe025995e", 00:23:54.830 "assigned_rate_limits": { 00:23:54.830 "rw_ios_per_sec": 0, 00:23:54.830 "rw_mbytes_per_sec": 0, 00:23:54.830 "r_mbytes_per_sec": 0, 00:23:54.830 "w_mbytes_per_sec": 0 00:23:54.830 }, 00:23:54.830 "claimed": true, 00:23:54.830 "claim_type": "exclusive_write", 00:23:54.830 "zoned": false, 00:23:54.830 "supported_io_types": { 00:23:54.830 "read": true, 00:23:54.830 "write": true, 00:23:54.830 "unmap": true, 00:23:54.830 "flush": true, 00:23:54.830 "reset": true, 00:23:54.830 "nvme_admin": false, 00:23:54.830 "nvme_io": false, 00:23:54.830 "nvme_io_md": false, 00:23:54.830 "write_zeroes": true, 00:23:54.830 "zcopy": true, 00:23:54.830 "get_zone_info": false, 00:23:54.830 "zone_management": false, 00:23:54.830 "zone_append": false, 00:23:54.830 "compare": false, 00:23:54.830 "compare_and_write": false, 00:23:54.830 "abort": true, 00:23:54.830 "seek_hole": false, 00:23:54.830 "seek_data": false, 00:23:54.830 "copy": true, 00:23:54.830 "nvme_iov_md": false 00:23:54.830 }, 00:23:54.830 "memory_domains": [ 00:23:54.830 { 00:23:54.830 "dma_device_id": "system", 00:23:54.830 "dma_device_type": 1 00:23:54.830 }, 00:23:54.830 { 00:23:54.830 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:54.830 "dma_device_type": 2 00:23:54.830 } 00:23:54.830 ], 00:23:54.830 "driver_specific": {} 00:23:54.830 } 00:23:54.830 ] 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:54.830 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:55.090 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:55.090 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:55.090 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:55.090 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:55.090 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:55.090 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:55.090 "name": "Existed_Raid", 00:23:55.090 "uuid": "8e49c8d0-f3d8-4493-89b5-ea9e4e4c3366", 00:23:55.090 "strip_size_kb": 64, 00:23:55.090 "state": "configuring", 00:23:55.090 "raid_level": "raid0", 00:23:55.090 "superblock": true, 00:23:55.090 "num_base_bdevs": 4, 00:23:55.090 "num_base_bdevs_discovered": 3, 00:23:55.090 "num_base_bdevs_operational": 4, 00:23:55.090 "base_bdevs_list": [ 00:23:55.090 { 00:23:55.090 "name": "BaseBdev1", 00:23:55.090 "uuid": "5b2f6391-eba3-4830-8ce9-c78fe025995e", 00:23:55.090 "is_configured": true, 00:23:55.090 "data_offset": 2048, 00:23:55.090 "data_size": 63488 00:23:55.090 }, 00:23:55.090 { 00:23:55.090 "name": null, 00:23:55.090 "uuid": "8f706d88-5059-42dd-b117-0b16828b2a07", 00:23:55.090 "is_configured": false, 00:23:55.090 "data_offset": 0, 00:23:55.090 "data_size": 63488 00:23:55.090 }, 00:23:55.090 { 00:23:55.090 "name": "BaseBdev3", 00:23:55.090 "uuid": "bd193d7a-a543-4daf-85df-a34a3b98a26a", 00:23:55.090 "is_configured": true, 00:23:55.090 "data_offset": 2048, 00:23:55.090 "data_size": 63488 00:23:55.090 }, 00:23:55.090 { 00:23:55.090 "name": "BaseBdev4", 00:23:55.090 "uuid": "104e730e-f926-4e5f-80c3-e17e1d838155", 00:23:55.090 "is_configured": true, 00:23:55.090 "data_offset": 2048, 00:23:55.090 "data_size": 63488 00:23:55.090 } 00:23:55.090 ] 00:23:55.090 }' 00:23:55.090 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:55.090 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:55.453 [2024-10-01 20:22:50.647285] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:55.453 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:55.714 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:55.714 "name": "Existed_Raid", 00:23:55.714 "uuid": "8e49c8d0-f3d8-4493-89b5-ea9e4e4c3366", 00:23:55.714 "strip_size_kb": 64, 00:23:55.714 "state": "configuring", 00:23:55.714 "raid_level": "raid0", 00:23:55.714 "superblock": true, 00:23:55.714 "num_base_bdevs": 4, 00:23:55.714 "num_base_bdevs_discovered": 2, 00:23:55.714 "num_base_bdevs_operational": 4, 00:23:55.714 "base_bdevs_list": [ 00:23:55.714 { 00:23:55.714 "name": "BaseBdev1", 00:23:55.714 "uuid": "5b2f6391-eba3-4830-8ce9-c78fe025995e", 00:23:55.714 "is_configured": true, 00:23:55.714 "data_offset": 2048, 00:23:55.714 "data_size": 63488 00:23:55.714 }, 00:23:55.714 { 00:23:55.714 "name": null, 00:23:55.714 "uuid": "8f706d88-5059-42dd-b117-0b16828b2a07", 00:23:55.714 "is_configured": false, 00:23:55.714 "data_offset": 0, 00:23:55.714 "data_size": 63488 00:23:55.714 }, 00:23:55.714 { 00:23:55.714 "name": null, 00:23:55.714 "uuid": "bd193d7a-a543-4daf-85df-a34a3b98a26a", 00:23:55.714 "is_configured": false, 00:23:55.714 "data_offset": 0, 00:23:55.714 "data_size": 63488 00:23:55.714 }, 00:23:55.714 { 00:23:55.714 "name": "BaseBdev4", 00:23:55.714 "uuid": "104e730e-f926-4e5f-80c3-e17e1d838155", 00:23:55.714 "is_configured": true, 00:23:55.714 "data_offset": 2048, 00:23:55.714 "data_size": 63488 00:23:55.714 } 00:23:55.714 ] 00:23:55.714 }' 00:23:55.714 20:22:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:55.714 20:22:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:55.974 [2024-10-01 20:22:51.195459] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:55.974 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:56.234 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:56.234 "name": "Existed_Raid", 00:23:56.234 "uuid": "8e49c8d0-f3d8-4493-89b5-ea9e4e4c3366", 00:23:56.234 "strip_size_kb": 64, 00:23:56.234 "state": "configuring", 00:23:56.234 "raid_level": "raid0", 00:23:56.234 "superblock": true, 00:23:56.234 "num_base_bdevs": 4, 00:23:56.234 "num_base_bdevs_discovered": 3, 00:23:56.234 "num_base_bdevs_operational": 4, 00:23:56.234 "base_bdevs_list": [ 00:23:56.234 { 00:23:56.234 "name": "BaseBdev1", 00:23:56.234 "uuid": "5b2f6391-eba3-4830-8ce9-c78fe025995e", 00:23:56.234 "is_configured": true, 00:23:56.234 "data_offset": 2048, 00:23:56.234 "data_size": 63488 00:23:56.234 }, 00:23:56.234 { 00:23:56.234 "name": null, 00:23:56.234 "uuid": "8f706d88-5059-42dd-b117-0b16828b2a07", 00:23:56.234 "is_configured": false, 00:23:56.234 "data_offset": 0, 00:23:56.234 "data_size": 63488 00:23:56.234 }, 00:23:56.234 { 00:23:56.234 "name": "BaseBdev3", 00:23:56.234 "uuid": "bd193d7a-a543-4daf-85df-a34a3b98a26a", 00:23:56.234 "is_configured": true, 00:23:56.234 "data_offset": 2048, 00:23:56.234 "data_size": 63488 00:23:56.234 }, 00:23:56.234 { 00:23:56.234 "name": "BaseBdev4", 00:23:56.234 "uuid": "104e730e-f926-4e5f-80c3-e17e1d838155", 00:23:56.234 "is_configured": true, 00:23:56.234 "data_offset": 2048, 00:23:56.234 "data_size": 63488 00:23:56.234 } 00:23:56.234 ] 00:23:56.234 }' 00:23:56.234 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:56.234 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:56.492 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:56.493 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:23:56.493 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:56.493 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:56.493 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:56.751 [2024-10-01 20:22:51.751599] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:56.751 "name": "Existed_Raid", 00:23:56.751 "uuid": "8e49c8d0-f3d8-4493-89b5-ea9e4e4c3366", 00:23:56.751 "strip_size_kb": 64, 00:23:56.751 "state": "configuring", 00:23:56.751 "raid_level": "raid0", 00:23:56.751 "superblock": true, 00:23:56.751 "num_base_bdevs": 4, 00:23:56.751 "num_base_bdevs_discovered": 2, 00:23:56.751 "num_base_bdevs_operational": 4, 00:23:56.751 "base_bdevs_list": [ 00:23:56.751 { 00:23:56.751 "name": null, 00:23:56.751 "uuid": "5b2f6391-eba3-4830-8ce9-c78fe025995e", 00:23:56.751 "is_configured": false, 00:23:56.751 "data_offset": 0, 00:23:56.751 "data_size": 63488 00:23:56.751 }, 00:23:56.751 { 00:23:56.751 "name": null, 00:23:56.751 "uuid": "8f706d88-5059-42dd-b117-0b16828b2a07", 00:23:56.751 "is_configured": false, 00:23:56.751 "data_offset": 0, 00:23:56.751 "data_size": 63488 00:23:56.751 }, 00:23:56.751 { 00:23:56.751 "name": "BaseBdev3", 00:23:56.751 "uuid": "bd193d7a-a543-4daf-85df-a34a3b98a26a", 00:23:56.751 "is_configured": true, 00:23:56.751 "data_offset": 2048, 00:23:56.751 "data_size": 63488 00:23:56.751 }, 00:23:56.751 { 00:23:56.751 "name": "BaseBdev4", 00:23:56.751 "uuid": "104e730e-f926-4e5f-80c3-e17e1d838155", 00:23:56.751 "is_configured": true, 00:23:56.751 "data_offset": 2048, 00:23:56.751 "data_size": 63488 00:23:56.751 } 00:23:56.751 ] 00:23:56.751 }' 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:56.751 20:22:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:57.319 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:57.320 [2024-10-01 20:22:52.392916] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:57.320 "name": "Existed_Raid", 00:23:57.320 "uuid": "8e49c8d0-f3d8-4493-89b5-ea9e4e4c3366", 00:23:57.320 "strip_size_kb": 64, 00:23:57.320 "state": "configuring", 00:23:57.320 "raid_level": "raid0", 00:23:57.320 "superblock": true, 00:23:57.320 "num_base_bdevs": 4, 00:23:57.320 "num_base_bdevs_discovered": 3, 00:23:57.320 "num_base_bdevs_operational": 4, 00:23:57.320 "base_bdevs_list": [ 00:23:57.320 { 00:23:57.320 "name": null, 00:23:57.320 "uuid": "5b2f6391-eba3-4830-8ce9-c78fe025995e", 00:23:57.320 "is_configured": false, 00:23:57.320 "data_offset": 0, 00:23:57.320 "data_size": 63488 00:23:57.320 }, 00:23:57.320 { 00:23:57.320 "name": "BaseBdev2", 00:23:57.320 "uuid": "8f706d88-5059-42dd-b117-0b16828b2a07", 00:23:57.320 "is_configured": true, 00:23:57.320 "data_offset": 2048, 00:23:57.320 "data_size": 63488 00:23:57.320 }, 00:23:57.320 { 00:23:57.320 "name": "BaseBdev3", 00:23:57.320 "uuid": "bd193d7a-a543-4daf-85df-a34a3b98a26a", 00:23:57.320 "is_configured": true, 00:23:57.320 "data_offset": 2048, 00:23:57.320 "data_size": 63488 00:23:57.320 }, 00:23:57.320 { 00:23:57.320 "name": "BaseBdev4", 00:23:57.320 "uuid": "104e730e-f926-4e5f-80c3-e17e1d838155", 00:23:57.320 "is_configured": true, 00:23:57.320 "data_offset": 2048, 00:23:57.320 "data_size": 63488 00:23:57.320 } 00:23:57.320 ] 00:23:57.320 }' 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:57.320 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:57.887 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:57.887 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:23:57.887 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:57.887 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:57.887 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:57.887 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:23:57.887 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:57.887 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:57.887 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:23:57.887 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:57.887 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:57.887 20:22:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 5b2f6391-eba3-4830-8ce9-c78fe025995e 00:23:57.887 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:57.887 20:22:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:57.887 [2024-10-01 20:22:53.023797] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:23:57.887 [2024-10-01 20:22:53.024112] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:23:57.887 [2024-10-01 20:22:53.024130] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:23:57.887 NewBaseBdev 00:23:57.887 [2024-10-01 20:22:53.024468] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:23:57.887 [2024-10-01 20:22:53.024635] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:23:57.887 [2024-10-01 20:22:53.024660] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:23:57.887 [2024-10-01 20:22:53.024842] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:23:57.887 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:57.887 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:23:57.887 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:23:57.887 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:23:57.887 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:23:57.887 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:23:57.887 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:23:57.887 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:23:57.887 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:57.887 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:57.887 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:57.887 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:23:57.887 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:57.887 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:57.887 [ 00:23:57.887 { 00:23:57.887 "name": "NewBaseBdev", 00:23:57.887 "aliases": [ 00:23:57.887 "5b2f6391-eba3-4830-8ce9-c78fe025995e" 00:23:57.887 ], 00:23:57.887 "product_name": "Malloc disk", 00:23:57.887 "block_size": 512, 00:23:57.887 "num_blocks": 65536, 00:23:57.888 "uuid": "5b2f6391-eba3-4830-8ce9-c78fe025995e", 00:23:57.888 "assigned_rate_limits": { 00:23:57.888 "rw_ios_per_sec": 0, 00:23:57.888 "rw_mbytes_per_sec": 0, 00:23:57.888 "r_mbytes_per_sec": 0, 00:23:57.888 "w_mbytes_per_sec": 0 00:23:57.888 }, 00:23:57.888 "claimed": true, 00:23:57.888 "claim_type": "exclusive_write", 00:23:57.888 "zoned": false, 00:23:57.888 "supported_io_types": { 00:23:57.888 "read": true, 00:23:57.888 "write": true, 00:23:57.888 "unmap": true, 00:23:57.888 "flush": true, 00:23:57.888 "reset": true, 00:23:57.888 "nvme_admin": false, 00:23:57.888 "nvme_io": false, 00:23:57.888 "nvme_io_md": false, 00:23:57.888 "write_zeroes": true, 00:23:57.888 "zcopy": true, 00:23:57.888 "get_zone_info": false, 00:23:57.888 "zone_management": false, 00:23:57.888 "zone_append": false, 00:23:57.888 "compare": false, 00:23:57.888 "compare_and_write": false, 00:23:57.888 "abort": true, 00:23:57.888 "seek_hole": false, 00:23:57.888 "seek_data": false, 00:23:57.888 "copy": true, 00:23:57.888 "nvme_iov_md": false 00:23:57.888 }, 00:23:57.888 "memory_domains": [ 00:23:57.888 { 00:23:57.888 "dma_device_id": "system", 00:23:57.888 "dma_device_type": 1 00:23:57.888 }, 00:23:57.888 { 00:23:57.888 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:57.888 "dma_device_type": 2 00:23:57.888 } 00:23:57.888 ], 00:23:57.888 "driver_specific": {} 00:23:57.888 } 00:23:57.888 ] 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:23:57.888 "name": "Existed_Raid", 00:23:57.888 "uuid": "8e49c8d0-f3d8-4493-89b5-ea9e4e4c3366", 00:23:57.888 "strip_size_kb": 64, 00:23:57.888 "state": "online", 00:23:57.888 "raid_level": "raid0", 00:23:57.888 "superblock": true, 00:23:57.888 "num_base_bdevs": 4, 00:23:57.888 "num_base_bdevs_discovered": 4, 00:23:57.888 "num_base_bdevs_operational": 4, 00:23:57.888 "base_bdevs_list": [ 00:23:57.888 { 00:23:57.888 "name": "NewBaseBdev", 00:23:57.888 "uuid": "5b2f6391-eba3-4830-8ce9-c78fe025995e", 00:23:57.888 "is_configured": true, 00:23:57.888 "data_offset": 2048, 00:23:57.888 "data_size": 63488 00:23:57.888 }, 00:23:57.888 { 00:23:57.888 "name": "BaseBdev2", 00:23:57.888 "uuid": "8f706d88-5059-42dd-b117-0b16828b2a07", 00:23:57.888 "is_configured": true, 00:23:57.888 "data_offset": 2048, 00:23:57.888 "data_size": 63488 00:23:57.888 }, 00:23:57.888 { 00:23:57.888 "name": "BaseBdev3", 00:23:57.888 "uuid": "bd193d7a-a543-4daf-85df-a34a3b98a26a", 00:23:57.888 "is_configured": true, 00:23:57.888 "data_offset": 2048, 00:23:57.888 "data_size": 63488 00:23:57.888 }, 00:23:57.888 { 00:23:57.888 "name": "BaseBdev4", 00:23:57.888 "uuid": "104e730e-f926-4e5f-80c3-e17e1d838155", 00:23:57.888 "is_configured": true, 00:23:57.888 "data_offset": 2048, 00:23:57.888 "data_size": 63488 00:23:57.888 } 00:23:57.888 ] 00:23:57.888 }' 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:23:57.888 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:58.456 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:23:58.456 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:23:58.456 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:23:58.456 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:23:58.456 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:23:58.456 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:23:58.456 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:23:58.456 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:58.456 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:58.456 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:23:58.456 [2024-10-01 20:22:53.572510] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:23:58.456 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:58.456 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:23:58.456 "name": "Existed_Raid", 00:23:58.456 "aliases": [ 00:23:58.456 "8e49c8d0-f3d8-4493-89b5-ea9e4e4c3366" 00:23:58.456 ], 00:23:58.456 "product_name": "Raid Volume", 00:23:58.456 "block_size": 512, 00:23:58.456 "num_blocks": 253952, 00:23:58.456 "uuid": "8e49c8d0-f3d8-4493-89b5-ea9e4e4c3366", 00:23:58.456 "assigned_rate_limits": { 00:23:58.456 "rw_ios_per_sec": 0, 00:23:58.456 "rw_mbytes_per_sec": 0, 00:23:58.456 "r_mbytes_per_sec": 0, 00:23:58.456 "w_mbytes_per_sec": 0 00:23:58.456 }, 00:23:58.456 "claimed": false, 00:23:58.456 "zoned": false, 00:23:58.456 "supported_io_types": { 00:23:58.456 "read": true, 00:23:58.456 "write": true, 00:23:58.456 "unmap": true, 00:23:58.456 "flush": true, 00:23:58.456 "reset": true, 00:23:58.456 "nvme_admin": false, 00:23:58.456 "nvme_io": false, 00:23:58.456 "nvme_io_md": false, 00:23:58.456 "write_zeroes": true, 00:23:58.456 "zcopy": false, 00:23:58.456 "get_zone_info": false, 00:23:58.456 "zone_management": false, 00:23:58.456 "zone_append": false, 00:23:58.456 "compare": false, 00:23:58.456 "compare_and_write": false, 00:23:58.456 "abort": false, 00:23:58.456 "seek_hole": false, 00:23:58.456 "seek_data": false, 00:23:58.456 "copy": false, 00:23:58.456 "nvme_iov_md": false 00:23:58.456 }, 00:23:58.456 "memory_domains": [ 00:23:58.456 { 00:23:58.456 "dma_device_id": "system", 00:23:58.456 "dma_device_type": 1 00:23:58.456 }, 00:23:58.456 { 00:23:58.456 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:58.456 "dma_device_type": 2 00:23:58.456 }, 00:23:58.456 { 00:23:58.456 "dma_device_id": "system", 00:23:58.456 "dma_device_type": 1 00:23:58.456 }, 00:23:58.456 { 00:23:58.456 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:58.456 "dma_device_type": 2 00:23:58.456 }, 00:23:58.456 { 00:23:58.456 "dma_device_id": "system", 00:23:58.456 "dma_device_type": 1 00:23:58.456 }, 00:23:58.456 { 00:23:58.456 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:58.456 "dma_device_type": 2 00:23:58.456 }, 00:23:58.456 { 00:23:58.456 "dma_device_id": "system", 00:23:58.456 "dma_device_type": 1 00:23:58.456 }, 00:23:58.456 { 00:23:58.456 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:23:58.456 "dma_device_type": 2 00:23:58.456 } 00:23:58.456 ], 00:23:58.456 "driver_specific": { 00:23:58.456 "raid": { 00:23:58.456 "uuid": "8e49c8d0-f3d8-4493-89b5-ea9e4e4c3366", 00:23:58.456 "strip_size_kb": 64, 00:23:58.456 "state": "online", 00:23:58.456 "raid_level": "raid0", 00:23:58.456 "superblock": true, 00:23:58.456 "num_base_bdevs": 4, 00:23:58.456 "num_base_bdevs_discovered": 4, 00:23:58.456 "num_base_bdevs_operational": 4, 00:23:58.456 "base_bdevs_list": [ 00:23:58.456 { 00:23:58.456 "name": "NewBaseBdev", 00:23:58.456 "uuid": "5b2f6391-eba3-4830-8ce9-c78fe025995e", 00:23:58.456 "is_configured": true, 00:23:58.456 "data_offset": 2048, 00:23:58.456 "data_size": 63488 00:23:58.456 }, 00:23:58.456 { 00:23:58.456 "name": "BaseBdev2", 00:23:58.456 "uuid": "8f706d88-5059-42dd-b117-0b16828b2a07", 00:23:58.456 "is_configured": true, 00:23:58.456 "data_offset": 2048, 00:23:58.456 "data_size": 63488 00:23:58.456 }, 00:23:58.456 { 00:23:58.456 "name": "BaseBdev3", 00:23:58.456 "uuid": "bd193d7a-a543-4daf-85df-a34a3b98a26a", 00:23:58.456 "is_configured": true, 00:23:58.456 "data_offset": 2048, 00:23:58.456 "data_size": 63488 00:23:58.456 }, 00:23:58.456 { 00:23:58.456 "name": "BaseBdev4", 00:23:58.456 "uuid": "104e730e-f926-4e5f-80c3-e17e1d838155", 00:23:58.456 "is_configured": true, 00:23:58.456 "data_offset": 2048, 00:23:58.456 "data_size": 63488 00:23:58.456 } 00:23:58.456 ] 00:23:58.456 } 00:23:58.456 } 00:23:58.456 }' 00:23:58.456 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:23:58.456 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:23:58.456 BaseBdev2 00:23:58.456 BaseBdev3 00:23:58.456 BaseBdev4' 00:23:58.456 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:23:58.715 [2024-10-01 20:22:53.936154] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:23:58.715 [2024-10-01 20:22:53.936194] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:23:58.715 [2024-10-01 20:22:53.936290] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:23:58.715 [2024-10-01 20:22:53.936384] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:23:58.715 [2024-10-01 20:22:53.936401] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:23:58.715 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:23:58.716 20:22:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 70783 00:23:58.716 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 70783 ']' 00:23:58.716 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 70783 00:23:58.716 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:23:58.716 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:58.716 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70783 00:23:58.974 killing process with pid 70783 00:23:58.974 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:23:58.974 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:23:58.974 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70783' 00:23:58.974 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 70783 00:23:58.974 [2024-10-01 20:22:53.987180] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:23:58.974 20:22:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 70783 00:23:59.233 [2024-10-01 20:22:54.342624] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:24:01.136 20:22:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:24:01.136 00:24:01.136 real 0m13.496s 00:24:01.136 user 0m21.708s 00:24:01.136 sys 0m1.914s 00:24:01.136 20:22:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:24:01.136 ************************************ 00:24:01.136 END TEST raid_state_function_test_sb 00:24:01.136 ************************************ 00:24:01.136 20:22:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:01.136 20:22:56 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 4 00:24:01.136 20:22:56 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:24:01.136 20:22:56 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:24:01.136 20:22:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:24:01.136 ************************************ 00:24:01.136 START TEST raid_superblock_test 00:24:01.136 ************************************ 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid0 4 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=71470 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 71470 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 71470 ']' 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:01.136 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:01.136 20:22:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:01.395 [2024-10-01 20:22:56.404023] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:24:01.395 [2024-10-01 20:22:56.404264] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71470 ] 00:24:01.395 [2024-10-01 20:22:56.587195] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:01.655 [2024-10-01 20:22:56.865086] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:24:01.914 [2024-10-01 20:22:57.071284] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:01.914 [2024-10-01 20:22:57.071364] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:02.483 malloc1 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:02.483 [2024-10-01 20:22:57.521632] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:24:02.483 [2024-10-01 20:22:57.521914] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:02.483 [2024-10-01 20:22:57.522006] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:24:02.483 [2024-10-01 20:22:57.522033] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:02.483 [2024-10-01 20:22:57.525285] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:02.483 [2024-10-01 20:22:57.525464] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:24:02.483 pt1 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:02.483 malloc2 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:02.483 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:02.483 [2024-10-01 20:22:57.579710] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:24:02.483 [2024-10-01 20:22:57.579956] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:02.483 [2024-10-01 20:22:57.580042] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:24:02.483 [2024-10-01 20:22:57.580250] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:02.483 [2024-10-01 20:22:57.583301] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:02.483 [2024-10-01 20:22:57.583485] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:24:02.483 pt2 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:02.484 malloc3 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:02.484 [2024-10-01 20:22:57.636445] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:24:02.484 [2024-10-01 20:22:57.636542] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:02.484 [2024-10-01 20:22:57.636576] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:24:02.484 [2024-10-01 20:22:57.636592] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:02.484 [2024-10-01 20:22:57.639550] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:02.484 [2024-10-01 20:22:57.639756] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:24:02.484 pt3 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:02.484 malloc4 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:02.484 [2024-10-01 20:22:57.692701] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:24:02.484 [2024-10-01 20:22:57.692940] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:02.484 [2024-10-01 20:22:57.692983] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:24:02.484 [2024-10-01 20:22:57.693000] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:02.484 [2024-10-01 20:22:57.695940] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:02.484 [2024-10-01 20:22:57.695985] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:24:02.484 pt4 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:02.484 [2024-10-01 20:22:57.704835] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:24:02.484 [2024-10-01 20:22:57.707430] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:24:02.484 [2024-10-01 20:22:57.707554] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:24:02.484 [2024-10-01 20:22:57.707665] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:24:02.484 [2024-10-01 20:22:57.707966] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:24:02.484 [2024-10-01 20:22:57.707993] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:24:02.484 [2024-10-01 20:22:57.708308] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:24:02.484 [2024-10-01 20:22:57.708530] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:24:02.484 [2024-10-01 20:22:57.708551] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:24:02.484 [2024-10-01 20:22:57.708787] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:02.484 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:02.743 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:02.743 "name": "raid_bdev1", 00:24:02.743 "uuid": "15c84570-0b55-4247-84d6-741cf4444e37", 00:24:02.743 "strip_size_kb": 64, 00:24:02.743 "state": "online", 00:24:02.743 "raid_level": "raid0", 00:24:02.743 "superblock": true, 00:24:02.743 "num_base_bdevs": 4, 00:24:02.743 "num_base_bdevs_discovered": 4, 00:24:02.743 "num_base_bdevs_operational": 4, 00:24:02.743 "base_bdevs_list": [ 00:24:02.743 { 00:24:02.743 "name": "pt1", 00:24:02.743 "uuid": "00000000-0000-0000-0000-000000000001", 00:24:02.743 "is_configured": true, 00:24:02.743 "data_offset": 2048, 00:24:02.743 "data_size": 63488 00:24:02.743 }, 00:24:02.743 { 00:24:02.743 "name": "pt2", 00:24:02.743 "uuid": "00000000-0000-0000-0000-000000000002", 00:24:02.743 "is_configured": true, 00:24:02.743 "data_offset": 2048, 00:24:02.743 "data_size": 63488 00:24:02.743 }, 00:24:02.743 { 00:24:02.743 "name": "pt3", 00:24:02.743 "uuid": "00000000-0000-0000-0000-000000000003", 00:24:02.743 "is_configured": true, 00:24:02.743 "data_offset": 2048, 00:24:02.743 "data_size": 63488 00:24:02.743 }, 00:24:02.743 { 00:24:02.743 "name": "pt4", 00:24:02.743 "uuid": "00000000-0000-0000-0000-000000000004", 00:24:02.743 "is_configured": true, 00:24:02.743 "data_offset": 2048, 00:24:02.743 "data_size": 63488 00:24:02.743 } 00:24:02.743 ] 00:24:02.743 }' 00:24:02.743 20:22:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:02.743 20:22:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.003 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:24:03.003 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:24:03.003 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:24:03.003 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:24:03.003 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:24:03.003 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:24:03.003 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:24:03.003 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.003 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:24:03.003 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.003 [2024-10-01 20:22:58.221491] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:24:03.003 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:03.262 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:03.262 "name": "raid_bdev1", 00:24:03.262 "aliases": [ 00:24:03.262 "15c84570-0b55-4247-84d6-741cf4444e37" 00:24:03.262 ], 00:24:03.262 "product_name": "Raid Volume", 00:24:03.262 "block_size": 512, 00:24:03.262 "num_blocks": 253952, 00:24:03.262 "uuid": "15c84570-0b55-4247-84d6-741cf4444e37", 00:24:03.262 "assigned_rate_limits": { 00:24:03.262 "rw_ios_per_sec": 0, 00:24:03.262 "rw_mbytes_per_sec": 0, 00:24:03.262 "r_mbytes_per_sec": 0, 00:24:03.262 "w_mbytes_per_sec": 0 00:24:03.262 }, 00:24:03.262 "claimed": false, 00:24:03.262 "zoned": false, 00:24:03.262 "supported_io_types": { 00:24:03.262 "read": true, 00:24:03.262 "write": true, 00:24:03.262 "unmap": true, 00:24:03.262 "flush": true, 00:24:03.262 "reset": true, 00:24:03.262 "nvme_admin": false, 00:24:03.262 "nvme_io": false, 00:24:03.262 "nvme_io_md": false, 00:24:03.262 "write_zeroes": true, 00:24:03.262 "zcopy": false, 00:24:03.262 "get_zone_info": false, 00:24:03.262 "zone_management": false, 00:24:03.262 "zone_append": false, 00:24:03.262 "compare": false, 00:24:03.262 "compare_and_write": false, 00:24:03.262 "abort": false, 00:24:03.262 "seek_hole": false, 00:24:03.262 "seek_data": false, 00:24:03.262 "copy": false, 00:24:03.262 "nvme_iov_md": false 00:24:03.262 }, 00:24:03.262 "memory_domains": [ 00:24:03.262 { 00:24:03.262 "dma_device_id": "system", 00:24:03.262 "dma_device_type": 1 00:24:03.262 }, 00:24:03.262 { 00:24:03.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:03.262 "dma_device_type": 2 00:24:03.262 }, 00:24:03.262 { 00:24:03.262 "dma_device_id": "system", 00:24:03.262 "dma_device_type": 1 00:24:03.262 }, 00:24:03.262 { 00:24:03.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:03.262 "dma_device_type": 2 00:24:03.262 }, 00:24:03.262 { 00:24:03.262 "dma_device_id": "system", 00:24:03.262 "dma_device_type": 1 00:24:03.262 }, 00:24:03.262 { 00:24:03.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:03.262 "dma_device_type": 2 00:24:03.262 }, 00:24:03.262 { 00:24:03.262 "dma_device_id": "system", 00:24:03.262 "dma_device_type": 1 00:24:03.262 }, 00:24:03.262 { 00:24:03.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:03.262 "dma_device_type": 2 00:24:03.262 } 00:24:03.262 ], 00:24:03.262 "driver_specific": { 00:24:03.262 "raid": { 00:24:03.262 "uuid": "15c84570-0b55-4247-84d6-741cf4444e37", 00:24:03.262 "strip_size_kb": 64, 00:24:03.262 "state": "online", 00:24:03.262 "raid_level": "raid0", 00:24:03.262 "superblock": true, 00:24:03.262 "num_base_bdevs": 4, 00:24:03.262 "num_base_bdevs_discovered": 4, 00:24:03.262 "num_base_bdevs_operational": 4, 00:24:03.262 "base_bdevs_list": [ 00:24:03.262 { 00:24:03.262 "name": "pt1", 00:24:03.262 "uuid": "00000000-0000-0000-0000-000000000001", 00:24:03.262 "is_configured": true, 00:24:03.262 "data_offset": 2048, 00:24:03.262 "data_size": 63488 00:24:03.262 }, 00:24:03.262 { 00:24:03.262 "name": "pt2", 00:24:03.262 "uuid": "00000000-0000-0000-0000-000000000002", 00:24:03.262 "is_configured": true, 00:24:03.262 "data_offset": 2048, 00:24:03.262 "data_size": 63488 00:24:03.262 }, 00:24:03.262 { 00:24:03.262 "name": "pt3", 00:24:03.262 "uuid": "00000000-0000-0000-0000-000000000003", 00:24:03.262 "is_configured": true, 00:24:03.262 "data_offset": 2048, 00:24:03.262 "data_size": 63488 00:24:03.262 }, 00:24:03.262 { 00:24:03.262 "name": "pt4", 00:24:03.263 "uuid": "00000000-0000-0000-0000-000000000004", 00:24:03.263 "is_configured": true, 00:24:03.263 "data_offset": 2048, 00:24:03.263 "data_size": 63488 00:24:03.263 } 00:24:03.263 ] 00:24:03.263 } 00:24:03.263 } 00:24:03.263 }' 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:24:03.263 pt2 00:24:03.263 pt3 00:24:03.263 pt4' 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:03.263 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.521 [2024-10-01 20:22:58.569557] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=15c84570-0b55-4247-84d6-741cf4444e37 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 15c84570-0b55-4247-84d6-741cf4444e37 ']' 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.521 [2024-10-01 20:22:58.621151] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:24:03.521 [2024-10-01 20:22:58.621182] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:24:03.521 [2024-10-01 20:22:58.621303] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:24:03.521 [2024-10-01 20:22:58.621405] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:24:03.521 [2024-10-01 20:22:58.621431] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.521 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:03.522 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:24:03.522 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:24:03.522 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.522 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.522 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:03.522 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:24:03.522 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:24:03.522 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.522 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.522 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:03.522 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:24:03.522 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.522 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.522 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:24:03.522 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:03.780 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:24:03.780 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:24:03.780 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:24:03.780 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:24:03.780 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:24:03.780 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:24:03.780 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:24:03.780 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.781 [2024-10-01 20:22:58.781190] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:24:03.781 [2024-10-01 20:22:58.783802] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:24:03.781 [2024-10-01 20:22:58.783872] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:24:03.781 [2024-10-01 20:22:58.783928] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:24:03.781 [2024-10-01 20:22:58.784003] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:24:03.781 [2024-10-01 20:22:58.784073] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:24:03.781 [2024-10-01 20:22:58.784108] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:24:03.781 [2024-10-01 20:22:58.784142] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:24:03.781 [2024-10-01 20:22:58.784166] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:24:03.781 [2024-10-01 20:22:58.784191] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:24:03.781 request: 00:24:03.781 { 00:24:03.781 "name": "raid_bdev1", 00:24:03.781 "raid_level": "raid0", 00:24:03.781 "base_bdevs": [ 00:24:03.781 "malloc1", 00:24:03.781 "malloc2", 00:24:03.781 "malloc3", 00:24:03.781 "malloc4" 00:24:03.781 ], 00:24:03.781 "strip_size_kb": 64, 00:24:03.781 "superblock": false, 00:24:03.781 "method": "bdev_raid_create", 00:24:03.781 "req_id": 1 00:24:03.781 } 00:24:03.781 Got JSON-RPC error response 00:24:03.781 response: 00:24:03.781 { 00:24:03.781 "code": -17, 00:24:03.781 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:24:03.781 } 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.781 [2024-10-01 20:22:58.845266] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:24:03.781 [2024-10-01 20:22:58.845488] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:03.781 [2024-10-01 20:22:58.845566] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:24:03.781 [2024-10-01 20:22:58.845770] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:03.781 [2024-10-01 20:22:58.848945] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:03.781 [2024-10-01 20:22:58.849153] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:24:03.781 [2024-10-01 20:22:58.849398] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:24:03.781 [2024-10-01 20:22:58.849592] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:24:03.781 pt1 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:03.781 "name": "raid_bdev1", 00:24:03.781 "uuid": "15c84570-0b55-4247-84d6-741cf4444e37", 00:24:03.781 "strip_size_kb": 64, 00:24:03.781 "state": "configuring", 00:24:03.781 "raid_level": "raid0", 00:24:03.781 "superblock": true, 00:24:03.781 "num_base_bdevs": 4, 00:24:03.781 "num_base_bdevs_discovered": 1, 00:24:03.781 "num_base_bdevs_operational": 4, 00:24:03.781 "base_bdevs_list": [ 00:24:03.781 { 00:24:03.781 "name": "pt1", 00:24:03.781 "uuid": "00000000-0000-0000-0000-000000000001", 00:24:03.781 "is_configured": true, 00:24:03.781 "data_offset": 2048, 00:24:03.781 "data_size": 63488 00:24:03.781 }, 00:24:03.781 { 00:24:03.781 "name": null, 00:24:03.781 "uuid": "00000000-0000-0000-0000-000000000002", 00:24:03.781 "is_configured": false, 00:24:03.781 "data_offset": 2048, 00:24:03.781 "data_size": 63488 00:24:03.781 }, 00:24:03.781 { 00:24:03.781 "name": null, 00:24:03.781 "uuid": "00000000-0000-0000-0000-000000000003", 00:24:03.781 "is_configured": false, 00:24:03.781 "data_offset": 2048, 00:24:03.781 "data_size": 63488 00:24:03.781 }, 00:24:03.781 { 00:24:03.781 "name": null, 00:24:03.781 "uuid": "00000000-0000-0000-0000-000000000004", 00:24:03.781 "is_configured": false, 00:24:03.781 "data_offset": 2048, 00:24:03.781 "data_size": 63488 00:24:03.781 } 00:24:03.781 ] 00:24:03.781 }' 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:03.781 20:22:58 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:04.347 [2024-10-01 20:22:59.393653] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:24:04.347 [2024-10-01 20:22:59.393910] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:04.347 [2024-10-01 20:22:59.393953] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:24:04.347 [2024-10-01 20:22:59.393973] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:04.347 [2024-10-01 20:22:59.394567] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:04.347 [2024-10-01 20:22:59.394605] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:24:04.347 [2024-10-01 20:22:59.394710] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:24:04.347 [2024-10-01 20:22:59.394765] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:24:04.347 pt2 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:04.347 [2024-10-01 20:22:59.405649] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:04.347 "name": "raid_bdev1", 00:24:04.347 "uuid": "15c84570-0b55-4247-84d6-741cf4444e37", 00:24:04.347 "strip_size_kb": 64, 00:24:04.347 "state": "configuring", 00:24:04.347 "raid_level": "raid0", 00:24:04.347 "superblock": true, 00:24:04.347 "num_base_bdevs": 4, 00:24:04.347 "num_base_bdevs_discovered": 1, 00:24:04.347 "num_base_bdevs_operational": 4, 00:24:04.347 "base_bdevs_list": [ 00:24:04.347 { 00:24:04.347 "name": "pt1", 00:24:04.347 "uuid": "00000000-0000-0000-0000-000000000001", 00:24:04.347 "is_configured": true, 00:24:04.347 "data_offset": 2048, 00:24:04.347 "data_size": 63488 00:24:04.347 }, 00:24:04.347 { 00:24:04.347 "name": null, 00:24:04.347 "uuid": "00000000-0000-0000-0000-000000000002", 00:24:04.347 "is_configured": false, 00:24:04.347 "data_offset": 0, 00:24:04.347 "data_size": 63488 00:24:04.347 }, 00:24:04.347 { 00:24:04.347 "name": null, 00:24:04.347 "uuid": "00000000-0000-0000-0000-000000000003", 00:24:04.347 "is_configured": false, 00:24:04.347 "data_offset": 2048, 00:24:04.347 "data_size": 63488 00:24:04.347 }, 00:24:04.347 { 00:24:04.347 "name": null, 00:24:04.347 "uuid": "00000000-0000-0000-0000-000000000004", 00:24:04.347 "is_configured": false, 00:24:04.347 "data_offset": 2048, 00:24:04.347 "data_size": 63488 00:24:04.347 } 00:24:04.347 ] 00:24:04.347 }' 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:04.347 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:04.914 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:24:04.914 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:24:04.914 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:24:04.914 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:04.914 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:04.914 [2024-10-01 20:22:59.909835] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:24:04.914 [2024-10-01 20:22:59.909919] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:04.915 [2024-10-01 20:22:59.909954] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:24:04.915 [2024-10-01 20:22:59.909970] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:04.915 [2024-10-01 20:22:59.910574] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:04.915 [2024-10-01 20:22:59.910599] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:24:04.915 [2024-10-01 20:22:59.910709] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:24:04.915 [2024-10-01 20:22:59.910779] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:24:04.915 pt2 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:04.915 [2024-10-01 20:22:59.917767] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:24:04.915 [2024-10-01 20:22:59.917825] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:04.915 [2024-10-01 20:22:59.917854] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:24:04.915 [2024-10-01 20:22:59.917868] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:04.915 [2024-10-01 20:22:59.918358] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:04.915 [2024-10-01 20:22:59.918390] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:24:04.915 [2024-10-01 20:22:59.918470] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:24:04.915 [2024-10-01 20:22:59.918504] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:24:04.915 pt3 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:04.915 [2024-10-01 20:22:59.925712] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:24:04.915 [2024-10-01 20:22:59.925953] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:04.915 [2024-10-01 20:22:59.925993] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:24:04.915 [2024-10-01 20:22:59.926009] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:04.915 [2024-10-01 20:22:59.926460] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:04.915 [2024-10-01 20:22:59.926485] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:24:04.915 [2024-10-01 20:22:59.926567] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:24:04.915 [2024-10-01 20:22:59.926594] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:24:04.915 [2024-10-01 20:22:59.926781] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:24:04.915 [2024-10-01 20:22:59.926799] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:24:04.915 [2024-10-01 20:22:59.927107] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:24:04.915 [2024-10-01 20:22:59.927293] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:24:04.915 [2024-10-01 20:22:59.927315] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:24:04.915 [2024-10-01 20:22:59.927471] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:04.915 pt4 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:04.915 "name": "raid_bdev1", 00:24:04.915 "uuid": "15c84570-0b55-4247-84d6-741cf4444e37", 00:24:04.915 "strip_size_kb": 64, 00:24:04.915 "state": "online", 00:24:04.915 "raid_level": "raid0", 00:24:04.915 "superblock": true, 00:24:04.915 "num_base_bdevs": 4, 00:24:04.915 "num_base_bdevs_discovered": 4, 00:24:04.915 "num_base_bdevs_operational": 4, 00:24:04.915 "base_bdevs_list": [ 00:24:04.915 { 00:24:04.915 "name": "pt1", 00:24:04.915 "uuid": "00000000-0000-0000-0000-000000000001", 00:24:04.915 "is_configured": true, 00:24:04.915 "data_offset": 2048, 00:24:04.915 "data_size": 63488 00:24:04.915 }, 00:24:04.915 { 00:24:04.915 "name": "pt2", 00:24:04.915 "uuid": "00000000-0000-0000-0000-000000000002", 00:24:04.915 "is_configured": true, 00:24:04.915 "data_offset": 2048, 00:24:04.915 "data_size": 63488 00:24:04.915 }, 00:24:04.915 { 00:24:04.915 "name": "pt3", 00:24:04.915 "uuid": "00000000-0000-0000-0000-000000000003", 00:24:04.915 "is_configured": true, 00:24:04.915 "data_offset": 2048, 00:24:04.915 "data_size": 63488 00:24:04.915 }, 00:24:04.915 { 00:24:04.915 "name": "pt4", 00:24:04.915 "uuid": "00000000-0000-0000-0000-000000000004", 00:24:04.915 "is_configured": true, 00:24:04.915 "data_offset": 2048, 00:24:04.915 "data_size": 63488 00:24:04.915 } 00:24:04.915 ] 00:24:04.915 }' 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:04.915 20:22:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:05.482 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:24:05.482 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:24:05.482 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:24:05.482 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:24:05.482 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:24:05.482 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:24:05.482 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:24:05.482 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:24:05.482 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:05.482 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:05.482 [2024-10-01 20:23:00.454364] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:24:05.482 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:05.482 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:05.482 "name": "raid_bdev1", 00:24:05.482 "aliases": [ 00:24:05.482 "15c84570-0b55-4247-84d6-741cf4444e37" 00:24:05.482 ], 00:24:05.482 "product_name": "Raid Volume", 00:24:05.482 "block_size": 512, 00:24:05.482 "num_blocks": 253952, 00:24:05.482 "uuid": "15c84570-0b55-4247-84d6-741cf4444e37", 00:24:05.482 "assigned_rate_limits": { 00:24:05.482 "rw_ios_per_sec": 0, 00:24:05.482 "rw_mbytes_per_sec": 0, 00:24:05.482 "r_mbytes_per_sec": 0, 00:24:05.482 "w_mbytes_per_sec": 0 00:24:05.482 }, 00:24:05.482 "claimed": false, 00:24:05.482 "zoned": false, 00:24:05.482 "supported_io_types": { 00:24:05.482 "read": true, 00:24:05.482 "write": true, 00:24:05.482 "unmap": true, 00:24:05.482 "flush": true, 00:24:05.482 "reset": true, 00:24:05.482 "nvme_admin": false, 00:24:05.482 "nvme_io": false, 00:24:05.482 "nvme_io_md": false, 00:24:05.482 "write_zeroes": true, 00:24:05.482 "zcopy": false, 00:24:05.482 "get_zone_info": false, 00:24:05.482 "zone_management": false, 00:24:05.482 "zone_append": false, 00:24:05.482 "compare": false, 00:24:05.482 "compare_and_write": false, 00:24:05.482 "abort": false, 00:24:05.482 "seek_hole": false, 00:24:05.482 "seek_data": false, 00:24:05.483 "copy": false, 00:24:05.483 "nvme_iov_md": false 00:24:05.483 }, 00:24:05.483 "memory_domains": [ 00:24:05.483 { 00:24:05.483 "dma_device_id": "system", 00:24:05.483 "dma_device_type": 1 00:24:05.483 }, 00:24:05.483 { 00:24:05.483 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:05.483 "dma_device_type": 2 00:24:05.483 }, 00:24:05.483 { 00:24:05.483 "dma_device_id": "system", 00:24:05.483 "dma_device_type": 1 00:24:05.483 }, 00:24:05.483 { 00:24:05.483 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:05.483 "dma_device_type": 2 00:24:05.483 }, 00:24:05.483 { 00:24:05.483 "dma_device_id": "system", 00:24:05.483 "dma_device_type": 1 00:24:05.483 }, 00:24:05.483 { 00:24:05.483 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:05.483 "dma_device_type": 2 00:24:05.483 }, 00:24:05.483 { 00:24:05.483 "dma_device_id": "system", 00:24:05.483 "dma_device_type": 1 00:24:05.483 }, 00:24:05.483 { 00:24:05.483 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:05.483 "dma_device_type": 2 00:24:05.483 } 00:24:05.483 ], 00:24:05.483 "driver_specific": { 00:24:05.483 "raid": { 00:24:05.483 "uuid": "15c84570-0b55-4247-84d6-741cf4444e37", 00:24:05.483 "strip_size_kb": 64, 00:24:05.483 "state": "online", 00:24:05.483 "raid_level": "raid0", 00:24:05.483 "superblock": true, 00:24:05.483 "num_base_bdevs": 4, 00:24:05.483 "num_base_bdevs_discovered": 4, 00:24:05.483 "num_base_bdevs_operational": 4, 00:24:05.483 "base_bdevs_list": [ 00:24:05.483 { 00:24:05.483 "name": "pt1", 00:24:05.483 "uuid": "00000000-0000-0000-0000-000000000001", 00:24:05.483 "is_configured": true, 00:24:05.483 "data_offset": 2048, 00:24:05.483 "data_size": 63488 00:24:05.483 }, 00:24:05.483 { 00:24:05.483 "name": "pt2", 00:24:05.483 "uuid": "00000000-0000-0000-0000-000000000002", 00:24:05.483 "is_configured": true, 00:24:05.483 "data_offset": 2048, 00:24:05.483 "data_size": 63488 00:24:05.483 }, 00:24:05.483 { 00:24:05.483 "name": "pt3", 00:24:05.483 "uuid": "00000000-0000-0000-0000-000000000003", 00:24:05.483 "is_configured": true, 00:24:05.483 "data_offset": 2048, 00:24:05.483 "data_size": 63488 00:24:05.483 }, 00:24:05.483 { 00:24:05.483 "name": "pt4", 00:24:05.483 "uuid": "00000000-0000-0000-0000-000000000004", 00:24:05.483 "is_configured": true, 00:24:05.483 "data_offset": 2048, 00:24:05.483 "data_size": 63488 00:24:05.483 } 00:24:05.483 ] 00:24:05.483 } 00:24:05.483 } 00:24:05.483 }' 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:24:05.483 pt2 00:24:05.483 pt3 00:24:05.483 pt4' 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:05.483 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:24:05.742 [2024-10-01 20:23:00.850410] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 15c84570-0b55-4247-84d6-741cf4444e37 '!=' 15c84570-0b55-4247-84d6-741cf4444e37 ']' 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 71470 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 71470 ']' 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 71470 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71470 00:24:05.742 killing process with pid 71470 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71470' 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 71470 00:24:05.742 [2024-10-01 20:23:00.932061] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:24:05.742 20:23:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 71470 00:24:05.742 [2024-10-01 20:23:00.932177] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:24:05.742 [2024-10-01 20:23:00.932279] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:24:05.742 [2024-10-01 20:23:00.932297] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:24:06.310 [2024-10-01 20:23:01.285676] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:24:08.212 ************************************ 00:24:08.212 END TEST raid_superblock_test 00:24:08.212 ************************************ 00:24:08.212 20:23:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:24:08.212 00:24:08.212 real 0m6.799s 00:24:08.212 user 0m9.648s 00:24:08.212 sys 0m1.095s 00:24:08.212 20:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:24:08.212 20:23:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:08.212 20:23:03 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 4 read 00:24:08.212 20:23:03 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:24:08.212 20:23:03 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:24:08.212 20:23:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:24:08.212 ************************************ 00:24:08.212 START TEST raid_read_error_test 00:24:08.212 ************************************ 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 4 read 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.mcWlB0zaSk 00:24:08.212 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=71748 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 71748 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 71748 ']' 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:08.212 20:23:03 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:08.212 [2024-10-01 20:23:03.198807] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:24:08.212 [2024-10-01 20:23:03.198965] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71748 ] 00:24:08.212 [2024-10-01 20:23:03.362858] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:08.501 [2024-10-01 20:23:03.595467] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:24:08.759 [2024-10-01 20:23:03.791465] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:08.759 [2024-10-01 20:23:03.791906] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:09.018 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:09.018 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:24:09.018 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:24:09.018 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:24:09.018 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:09.018 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:09.019 BaseBdev1_malloc 00:24:09.019 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:09.019 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:24:09.019 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:09.019 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:09.019 true 00:24:09.019 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:09.019 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:24:09.019 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:09.019 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:09.019 [2024-10-01 20:23:04.267887] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:24:09.019 [2024-10-01 20:23:04.267954] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:09.019 [2024-10-01 20:23:04.267983] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:24:09.019 [2024-10-01 20:23:04.268002] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:09.278 [2024-10-01 20:23:04.271841] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:09.278 [2024-10-01 20:23:04.271898] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:24:09.278 BaseBdev1 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:09.278 BaseBdev2_malloc 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:09.278 true 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:09.278 [2024-10-01 20:23:04.333410] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:24:09.278 [2024-10-01 20:23:04.333685] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:09.278 [2024-10-01 20:23:04.333799] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:24:09.278 [2024-10-01 20:23:04.334097] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:09.278 [2024-10-01 20:23:04.337365] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:09.278 [2024-10-01 20:23:04.337530] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:24:09.278 BaseBdev2 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:09.278 BaseBdev3_malloc 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:09.278 true 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:09.278 [2024-10-01 20:23:04.399024] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:24:09.278 [2024-10-01 20:23:04.399113] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:09.278 [2024-10-01 20:23:04.399143] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:24:09.278 [2024-10-01 20:23:04.399162] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:09.278 [2024-10-01 20:23:04.402183] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:09.278 [2024-10-01 20:23:04.402235] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:24:09.278 BaseBdev3 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:09.278 BaseBdev4_malloc 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:09.278 true 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:09.278 [2024-10-01 20:23:04.456380] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:24:09.278 [2024-10-01 20:23:04.456566] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:09.278 [2024-10-01 20:23:04.456605] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:24:09.278 [2024-10-01 20:23:04.456626] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:09.278 [2024-10-01 20:23:04.459575] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:09.278 [2024-10-01 20:23:04.459763] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:24:09.278 BaseBdev4 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:09.278 [2024-10-01 20:23:04.464625] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:24:09.278 [2024-10-01 20:23:04.467173] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:24:09.278 [2024-10-01 20:23:04.467272] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:24:09.278 [2024-10-01 20:23:04.467363] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:24:09.278 [2024-10-01 20:23:04.467694] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:24:09.278 [2024-10-01 20:23:04.467761] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:24:09.278 [2024-10-01 20:23:04.468083] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:24:09.278 [2024-10-01 20:23:04.468309] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:24:09.278 [2024-10-01 20:23:04.468332] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:24:09.278 [2024-10-01 20:23:04.468589] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:09.278 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:09.537 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:09.537 "name": "raid_bdev1", 00:24:09.537 "uuid": "5fecc496-aac5-44ce-a91c-b57955b0d2b5", 00:24:09.537 "strip_size_kb": 64, 00:24:09.537 "state": "online", 00:24:09.537 "raid_level": "raid0", 00:24:09.537 "superblock": true, 00:24:09.537 "num_base_bdevs": 4, 00:24:09.537 "num_base_bdevs_discovered": 4, 00:24:09.537 "num_base_bdevs_operational": 4, 00:24:09.537 "base_bdevs_list": [ 00:24:09.537 { 00:24:09.537 "name": "BaseBdev1", 00:24:09.537 "uuid": "64b2e469-b909-545e-8e30-1f2bfbcf1e23", 00:24:09.537 "is_configured": true, 00:24:09.537 "data_offset": 2048, 00:24:09.537 "data_size": 63488 00:24:09.537 }, 00:24:09.537 { 00:24:09.537 "name": "BaseBdev2", 00:24:09.537 "uuid": "722bfc6c-52a0-56c9-bee6-805cb9c4b1d8", 00:24:09.537 "is_configured": true, 00:24:09.537 "data_offset": 2048, 00:24:09.537 "data_size": 63488 00:24:09.537 }, 00:24:09.537 { 00:24:09.537 "name": "BaseBdev3", 00:24:09.537 "uuid": "27a70717-2ccb-55ee-b65b-f194a136d899", 00:24:09.537 "is_configured": true, 00:24:09.537 "data_offset": 2048, 00:24:09.537 "data_size": 63488 00:24:09.537 }, 00:24:09.537 { 00:24:09.537 "name": "BaseBdev4", 00:24:09.537 "uuid": "4303a045-7787-5450-a459-a94cff6f0c9c", 00:24:09.537 "is_configured": true, 00:24:09.537 "data_offset": 2048, 00:24:09.537 "data_size": 63488 00:24:09.537 } 00:24:09.537 ] 00:24:09.537 }' 00:24:09.537 20:23:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:09.537 20:23:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:09.795 20:23:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:24:09.795 20:23:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:24:10.053 [2024-10-01 20:23:05.114419] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:10.987 "name": "raid_bdev1", 00:24:10.987 "uuid": "5fecc496-aac5-44ce-a91c-b57955b0d2b5", 00:24:10.987 "strip_size_kb": 64, 00:24:10.987 "state": "online", 00:24:10.987 "raid_level": "raid0", 00:24:10.987 "superblock": true, 00:24:10.987 "num_base_bdevs": 4, 00:24:10.987 "num_base_bdevs_discovered": 4, 00:24:10.987 "num_base_bdevs_operational": 4, 00:24:10.987 "base_bdevs_list": [ 00:24:10.987 { 00:24:10.987 "name": "BaseBdev1", 00:24:10.987 "uuid": "64b2e469-b909-545e-8e30-1f2bfbcf1e23", 00:24:10.987 "is_configured": true, 00:24:10.987 "data_offset": 2048, 00:24:10.987 "data_size": 63488 00:24:10.987 }, 00:24:10.987 { 00:24:10.987 "name": "BaseBdev2", 00:24:10.987 "uuid": "722bfc6c-52a0-56c9-bee6-805cb9c4b1d8", 00:24:10.987 "is_configured": true, 00:24:10.987 "data_offset": 2048, 00:24:10.987 "data_size": 63488 00:24:10.987 }, 00:24:10.987 { 00:24:10.987 "name": "BaseBdev3", 00:24:10.987 "uuid": "27a70717-2ccb-55ee-b65b-f194a136d899", 00:24:10.987 "is_configured": true, 00:24:10.987 "data_offset": 2048, 00:24:10.987 "data_size": 63488 00:24:10.987 }, 00:24:10.987 { 00:24:10.987 "name": "BaseBdev4", 00:24:10.987 "uuid": "4303a045-7787-5450-a459-a94cff6f0c9c", 00:24:10.987 "is_configured": true, 00:24:10.987 "data_offset": 2048, 00:24:10.987 "data_size": 63488 00:24:10.987 } 00:24:10.987 ] 00:24:10.987 }' 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:10.987 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:11.553 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:24:11.553 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:11.553 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:11.553 [2024-10-01 20:23:06.545023] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:24:11.553 [2024-10-01 20:23:06.545255] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:24:11.553 [2024-10-01 20:23:06.548762] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:24:11.553 [2024-10-01 20:23:06.548848] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:11.553 [2024-10-01 20:23:06.548915] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:24:11.553 [2024-10-01 20:23:06.548935] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:24:11.553 { 00:24:11.553 "results": [ 00:24:11.553 { 00:24:11.553 "job": "raid_bdev1", 00:24:11.553 "core_mask": "0x1", 00:24:11.553 "workload": "randrw", 00:24:11.553 "percentage": 50, 00:24:11.553 "status": "finished", 00:24:11.553 "queue_depth": 1, 00:24:11.553 "io_size": 131072, 00:24:11.553 "runtime": 1.428041, 00:24:11.553 "iops": 9822.54711174259, 00:24:11.553 "mibps": 1227.8183889678237, 00:24:11.553 "io_failed": 1, 00:24:11.553 "io_timeout": 0, 00:24:11.553 "avg_latency_us": 142.2452964201467, 00:24:11.553 "min_latency_us": 41.42545454545454, 00:24:11.553 "max_latency_us": 1884.16 00:24:11.553 } 00:24:11.553 ], 00:24:11.553 "core_count": 1 00:24:11.553 } 00:24:11.553 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:11.553 20:23:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 71748 00:24:11.553 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 71748 ']' 00:24:11.554 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 71748 00:24:11.554 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:24:11.554 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:11.554 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71748 00:24:11.554 killing process with pid 71748 00:24:11.554 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:24:11.554 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:24:11.554 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71748' 00:24:11.554 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 71748 00:24:11.554 [2024-10-01 20:23:06.579937] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:24:11.554 20:23:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 71748 00:24:11.813 [2024-10-01 20:23:06.876971] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:24:13.716 20:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:24:13.716 20:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.mcWlB0zaSk 00:24:13.716 20:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:24:13.716 ************************************ 00:24:13.716 END TEST raid_read_error_test 00:24:13.716 ************************************ 00:24:13.716 20:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.70 00:24:13.716 20:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:24:13.716 20:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:24:13.716 20:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:24:13.716 20:23:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.70 != \0\.\0\0 ]] 00:24:13.716 00:24:13.716 real 0m5.686s 00:24:13.716 user 0m6.686s 00:24:13.716 sys 0m0.729s 00:24:13.716 20:23:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:24:13.716 20:23:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:13.716 20:23:08 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 4 write 00:24:13.716 20:23:08 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:24:13.716 20:23:08 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:24:13.716 20:23:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:24:13.716 ************************************ 00:24:13.716 START TEST raid_write_error_test 00:24:13.716 ************************************ 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 4 write 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.0QYvRlfPlT 00:24:13.716 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=71903 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 71903 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 71903 ']' 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:13.716 20:23:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:13.716 [2024-10-01 20:23:08.952410] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:24:13.716 [2024-10-01 20:23:08.952902] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71903 ] 00:24:13.975 [2024-10-01 20:23:09.135770] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:14.235 [2024-10-01 20:23:09.414930] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:24:14.494 [2024-10-01 20:23:09.640054] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:14.494 [2024-10-01 20:23:09.640169] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:15.061 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:15.061 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:24:15.061 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:24:15.061 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:24:15.061 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:15.061 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:15.061 BaseBdev1_malloc 00:24:15.061 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:15.062 true 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:15.062 [2024-10-01 20:23:10.119131] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:24:15.062 [2024-10-01 20:23:10.119225] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:15.062 [2024-10-01 20:23:10.119287] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:24:15.062 [2024-10-01 20:23:10.119306] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:15.062 [2024-10-01 20:23:10.122386] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:15.062 [2024-10-01 20:23:10.122451] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:24:15.062 BaseBdev1 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:15.062 BaseBdev2_malloc 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:15.062 true 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:15.062 [2024-10-01 20:23:10.185606] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:24:15.062 [2024-10-01 20:23:10.185855] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:15.062 [2024-10-01 20:23:10.185930] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:24:15.062 [2024-10-01 20:23:10.186055] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:15.062 [2024-10-01 20:23:10.189135] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:15.062 BaseBdev2 00:24:15.062 [2024-10-01 20:23:10.189308] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:15.062 BaseBdev3_malloc 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:15.062 true 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:15.062 [2024-10-01 20:23:10.242939] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:24:15.062 [2024-10-01 20:23:10.243141] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:15.062 [2024-10-01 20:23:10.243184] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:24:15.062 [2024-10-01 20:23:10.243206] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:15.062 [2024-10-01 20:23:10.246264] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:15.062 [2024-10-01 20:23:10.246333] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:24:15.062 BaseBdev3 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:15.062 BaseBdev4_malloc 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:15.062 true 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:15.062 [2024-10-01 20:23:10.300605] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:24:15.062 [2024-10-01 20:23:10.300844] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:15.062 [2024-10-01 20:23:10.300887] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:24:15.062 [2024-10-01 20:23:10.300912] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:15.062 [2024-10-01 20:23:10.303955] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:15.062 [2024-10-01 20:23:10.304023] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:24:15.062 BaseBdev4 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:15.062 [2024-10-01 20:23:10.308767] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:24:15.062 [2024-10-01 20:23:10.311375] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:24:15.062 [2024-10-01 20:23:10.311487] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:24:15.062 [2024-10-01 20:23:10.311593] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:24:15.062 [2024-10-01 20:23:10.311942] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:24:15.062 [2024-10-01 20:23:10.312024] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:24:15.062 [2024-10-01 20:23:10.312391] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:24:15.062 [2024-10-01 20:23:10.312615] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:24:15.062 [2024-10-01 20:23:10.312631] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:24:15.062 [2024-10-01 20:23:10.312942] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:15.062 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:15.321 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:24:15.321 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:24:15.321 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:24:15.321 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:24:15.321 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:15.321 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:15.321 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:15.321 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:15.321 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:15.322 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:15.322 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:15.322 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:15.322 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:15.322 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:15.322 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:15.322 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:15.322 "name": "raid_bdev1", 00:24:15.322 "uuid": "6cb0f196-3222-4361-b2d4-a8ed3ed4e408", 00:24:15.322 "strip_size_kb": 64, 00:24:15.322 "state": "online", 00:24:15.322 "raid_level": "raid0", 00:24:15.322 "superblock": true, 00:24:15.322 "num_base_bdevs": 4, 00:24:15.322 "num_base_bdevs_discovered": 4, 00:24:15.322 "num_base_bdevs_operational": 4, 00:24:15.322 "base_bdevs_list": [ 00:24:15.322 { 00:24:15.322 "name": "BaseBdev1", 00:24:15.322 "uuid": "81baaf44-36e3-51e4-a04f-1fc7a0ec3606", 00:24:15.322 "is_configured": true, 00:24:15.322 "data_offset": 2048, 00:24:15.322 "data_size": 63488 00:24:15.322 }, 00:24:15.322 { 00:24:15.322 "name": "BaseBdev2", 00:24:15.322 "uuid": "ee6b5fd6-0e1d-5fc9-bafd-a8a6cc9193d7", 00:24:15.322 "is_configured": true, 00:24:15.322 "data_offset": 2048, 00:24:15.322 "data_size": 63488 00:24:15.322 }, 00:24:15.322 { 00:24:15.322 "name": "BaseBdev3", 00:24:15.322 "uuid": "f16e7fad-26ba-5e9b-8c77-c5ee41bd911f", 00:24:15.322 "is_configured": true, 00:24:15.322 "data_offset": 2048, 00:24:15.322 "data_size": 63488 00:24:15.322 }, 00:24:15.322 { 00:24:15.322 "name": "BaseBdev4", 00:24:15.322 "uuid": "8fd3fc08-24d9-5dba-b4e3-d93471b5274b", 00:24:15.322 "is_configured": true, 00:24:15.322 "data_offset": 2048, 00:24:15.322 "data_size": 63488 00:24:15.322 } 00:24:15.322 ] 00:24:15.322 }' 00:24:15.322 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:15.322 20:23:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:15.891 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:24:15.891 20:23:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:24:15.891 [2024-10-01 20:23:10.990661] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:16.828 "name": "raid_bdev1", 00:24:16.828 "uuid": "6cb0f196-3222-4361-b2d4-a8ed3ed4e408", 00:24:16.828 "strip_size_kb": 64, 00:24:16.828 "state": "online", 00:24:16.828 "raid_level": "raid0", 00:24:16.828 "superblock": true, 00:24:16.828 "num_base_bdevs": 4, 00:24:16.828 "num_base_bdevs_discovered": 4, 00:24:16.828 "num_base_bdevs_operational": 4, 00:24:16.828 "base_bdevs_list": [ 00:24:16.828 { 00:24:16.828 "name": "BaseBdev1", 00:24:16.828 "uuid": "81baaf44-36e3-51e4-a04f-1fc7a0ec3606", 00:24:16.828 "is_configured": true, 00:24:16.828 "data_offset": 2048, 00:24:16.828 "data_size": 63488 00:24:16.828 }, 00:24:16.828 { 00:24:16.828 "name": "BaseBdev2", 00:24:16.828 "uuid": "ee6b5fd6-0e1d-5fc9-bafd-a8a6cc9193d7", 00:24:16.828 "is_configured": true, 00:24:16.828 "data_offset": 2048, 00:24:16.828 "data_size": 63488 00:24:16.828 }, 00:24:16.828 { 00:24:16.828 "name": "BaseBdev3", 00:24:16.828 "uuid": "f16e7fad-26ba-5e9b-8c77-c5ee41bd911f", 00:24:16.828 "is_configured": true, 00:24:16.828 "data_offset": 2048, 00:24:16.828 "data_size": 63488 00:24:16.828 }, 00:24:16.828 { 00:24:16.828 "name": "BaseBdev4", 00:24:16.828 "uuid": "8fd3fc08-24d9-5dba-b4e3-d93471b5274b", 00:24:16.828 "is_configured": true, 00:24:16.828 "data_offset": 2048, 00:24:16.828 "data_size": 63488 00:24:16.828 } 00:24:16.828 ] 00:24:16.828 }' 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:16.828 20:23:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:17.397 20:23:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:24:17.397 20:23:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:17.397 20:23:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:17.397 [2024-10-01 20:23:12.369072] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:24:17.397 [2024-10-01 20:23:12.369115] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:24:17.397 [2024-10-01 20:23:12.372776] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:24:17.397 [2024-10-01 20:23:12.372974] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:17.397 [2024-10-01 20:23:12.373083] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:24:17.397 [2024-10-01 20:23:12.373315] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:24:17.397 20:23:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:17.397 { 00:24:17.397 "results": [ 00:24:17.397 { 00:24:17.397 "job": "raid_bdev1", 00:24:17.397 "core_mask": "0x1", 00:24:17.397 "workload": "randrw", 00:24:17.397 "percentage": 50, 00:24:17.397 "status": "finished", 00:24:17.397 "queue_depth": 1, 00:24:17.397 "io_size": 131072, 00:24:17.397 "runtime": 1.375476, 00:24:17.397 "iops": 9754.441371568824, 00:24:17.397 "mibps": 1219.305171446103, 00:24:17.397 "io_failed": 1, 00:24:17.397 "io_timeout": 0, 00:24:17.397 "avg_latency_us": 142.98092060868035, 00:24:17.397 "min_latency_us": 40.72727272727273, 00:24:17.397 "max_latency_us": 1951.1854545454546 00:24:17.397 } 00:24:17.397 ], 00:24:17.397 "core_count": 1 00:24:17.397 } 00:24:17.397 20:23:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 71903 00:24:17.397 20:23:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 71903 ']' 00:24:17.397 20:23:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 71903 00:24:17.397 20:23:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:24:17.397 20:23:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:17.398 20:23:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71903 00:24:17.398 killing process with pid 71903 00:24:17.398 20:23:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:24:17.398 20:23:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:24:17.398 20:23:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71903' 00:24:17.398 20:23:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 71903 00:24:17.398 20:23:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 71903 00:24:17.398 [2024-10-01 20:23:12.406414] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:24:17.657 [2024-10-01 20:23:12.708464] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:24:19.562 20:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.0QYvRlfPlT 00:24:19.562 20:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:24:19.562 20:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:24:19.562 20:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:24:19.562 20:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:24:19.562 20:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:24:19.562 20:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:24:19.562 20:23:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:24:19.562 00:24:19.562 real 0m5.790s 00:24:19.562 user 0m6.800s 00:24:19.562 sys 0m0.778s 00:24:19.562 20:23:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:24:19.562 20:23:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:19.562 ************************************ 00:24:19.562 END TEST raid_write_error_test 00:24:19.562 ************************************ 00:24:19.562 20:23:14 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:24:19.562 20:23:14 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 4 false 00:24:19.562 20:23:14 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:24:19.562 20:23:14 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:24:19.562 20:23:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:24:19.562 ************************************ 00:24:19.562 START TEST raid_state_function_test 00:24:19.562 ************************************ 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 4 false 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:24:19.563 Process raid pid: 72059 00:24:19.563 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=72059 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72059' 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 72059 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 72059 ']' 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:19.563 20:23:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:19.563 [2024-10-01 20:23:14.786429] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:24:19.563 [2024-10-01 20:23:14.786916] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:19.822 [2024-10-01 20:23:14.967661] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:20.081 [2024-10-01 20:23:15.235505] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:24:20.339 [2024-10-01 20:23:15.441255] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:20.339 [2024-10-01 20:23:15.441550] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:20.908 [2024-10-01 20:23:15.864902] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:24:20.908 [2024-10-01 20:23:15.864981] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:24:20.908 [2024-10-01 20:23:15.864999] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:24:20.908 [2024-10-01 20:23:15.865017] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:24:20.908 [2024-10-01 20:23:15.865028] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:24:20.908 [2024-10-01 20:23:15.865046] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:24:20.908 [2024-10-01 20:23:15.865057] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:24:20.908 [2024-10-01 20:23:15.865072] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:20.908 "name": "Existed_Raid", 00:24:20.908 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:20.908 "strip_size_kb": 64, 00:24:20.908 "state": "configuring", 00:24:20.908 "raid_level": "concat", 00:24:20.908 "superblock": false, 00:24:20.908 "num_base_bdevs": 4, 00:24:20.908 "num_base_bdevs_discovered": 0, 00:24:20.908 "num_base_bdevs_operational": 4, 00:24:20.908 "base_bdevs_list": [ 00:24:20.908 { 00:24:20.908 "name": "BaseBdev1", 00:24:20.908 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:20.908 "is_configured": false, 00:24:20.908 "data_offset": 0, 00:24:20.908 "data_size": 0 00:24:20.908 }, 00:24:20.908 { 00:24:20.908 "name": "BaseBdev2", 00:24:20.908 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:20.908 "is_configured": false, 00:24:20.908 "data_offset": 0, 00:24:20.908 "data_size": 0 00:24:20.908 }, 00:24:20.908 { 00:24:20.908 "name": "BaseBdev3", 00:24:20.908 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:20.908 "is_configured": false, 00:24:20.908 "data_offset": 0, 00:24:20.908 "data_size": 0 00:24:20.908 }, 00:24:20.908 { 00:24:20.908 "name": "BaseBdev4", 00:24:20.908 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:20.908 "is_configured": false, 00:24:20.908 "data_offset": 0, 00:24:20.908 "data_size": 0 00:24:20.908 } 00:24:20.908 ] 00:24:20.908 }' 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:20.908 20:23:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:21.167 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:24:21.167 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:21.167 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:21.426 [2024-10-01 20:23:16.424985] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:24:21.426 [2024-10-01 20:23:16.425187] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:21.426 [2024-10-01 20:23:16.432991] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:24:21.426 [2024-10-01 20:23:16.433189] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:24:21.426 [2024-10-01 20:23:16.433347] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:24:21.426 [2024-10-01 20:23:16.433407] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:24:21.426 [2024-10-01 20:23:16.433639] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:24:21.426 [2024-10-01 20:23:16.433703] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:24:21.426 [2024-10-01 20:23:16.433920] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:24:21.426 [2024-10-01 20:23:16.433981] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:21.426 [2024-10-01 20:23:16.479973] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:24:21.426 BaseBdev1 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:21.426 [ 00:24:21.426 { 00:24:21.426 "name": "BaseBdev1", 00:24:21.426 "aliases": [ 00:24:21.426 "7684bd78-5c0f-430d-89a6-7bbd2b151bbd" 00:24:21.426 ], 00:24:21.426 "product_name": "Malloc disk", 00:24:21.426 "block_size": 512, 00:24:21.426 "num_blocks": 65536, 00:24:21.426 "uuid": "7684bd78-5c0f-430d-89a6-7bbd2b151bbd", 00:24:21.426 "assigned_rate_limits": { 00:24:21.426 "rw_ios_per_sec": 0, 00:24:21.426 "rw_mbytes_per_sec": 0, 00:24:21.426 "r_mbytes_per_sec": 0, 00:24:21.426 "w_mbytes_per_sec": 0 00:24:21.426 }, 00:24:21.426 "claimed": true, 00:24:21.426 "claim_type": "exclusive_write", 00:24:21.426 "zoned": false, 00:24:21.426 "supported_io_types": { 00:24:21.426 "read": true, 00:24:21.426 "write": true, 00:24:21.426 "unmap": true, 00:24:21.426 "flush": true, 00:24:21.426 "reset": true, 00:24:21.426 "nvme_admin": false, 00:24:21.426 "nvme_io": false, 00:24:21.426 "nvme_io_md": false, 00:24:21.426 "write_zeroes": true, 00:24:21.426 "zcopy": true, 00:24:21.426 "get_zone_info": false, 00:24:21.426 "zone_management": false, 00:24:21.426 "zone_append": false, 00:24:21.426 "compare": false, 00:24:21.426 "compare_and_write": false, 00:24:21.426 "abort": true, 00:24:21.426 "seek_hole": false, 00:24:21.426 "seek_data": false, 00:24:21.426 "copy": true, 00:24:21.426 "nvme_iov_md": false 00:24:21.426 }, 00:24:21.426 "memory_domains": [ 00:24:21.426 { 00:24:21.426 "dma_device_id": "system", 00:24:21.426 "dma_device_type": 1 00:24:21.426 }, 00:24:21.426 { 00:24:21.426 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:21.426 "dma_device_type": 2 00:24:21.426 } 00:24:21.426 ], 00:24:21.426 "driver_specific": {} 00:24:21.426 } 00:24:21.426 ] 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:21.426 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:21.427 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:21.427 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:21.427 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:21.427 "name": "Existed_Raid", 00:24:21.427 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:21.427 "strip_size_kb": 64, 00:24:21.427 "state": "configuring", 00:24:21.427 "raid_level": "concat", 00:24:21.427 "superblock": false, 00:24:21.427 "num_base_bdevs": 4, 00:24:21.427 "num_base_bdevs_discovered": 1, 00:24:21.427 "num_base_bdevs_operational": 4, 00:24:21.427 "base_bdevs_list": [ 00:24:21.427 { 00:24:21.427 "name": "BaseBdev1", 00:24:21.427 "uuid": "7684bd78-5c0f-430d-89a6-7bbd2b151bbd", 00:24:21.427 "is_configured": true, 00:24:21.427 "data_offset": 0, 00:24:21.427 "data_size": 65536 00:24:21.427 }, 00:24:21.427 { 00:24:21.427 "name": "BaseBdev2", 00:24:21.427 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:21.427 "is_configured": false, 00:24:21.427 "data_offset": 0, 00:24:21.427 "data_size": 0 00:24:21.427 }, 00:24:21.427 { 00:24:21.427 "name": "BaseBdev3", 00:24:21.427 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:21.427 "is_configured": false, 00:24:21.427 "data_offset": 0, 00:24:21.427 "data_size": 0 00:24:21.427 }, 00:24:21.427 { 00:24:21.427 "name": "BaseBdev4", 00:24:21.427 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:21.427 "is_configured": false, 00:24:21.427 "data_offset": 0, 00:24:21.427 "data_size": 0 00:24:21.427 } 00:24:21.427 ] 00:24:21.427 }' 00:24:21.427 20:23:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:21.427 20:23:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:22.017 [2024-10-01 20:23:17.044212] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:24:22.017 [2024-10-01 20:23:17.044293] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:22.017 [2024-10-01 20:23:17.052302] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:24:22.017 [2024-10-01 20:23:17.055221] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:24:22.017 [2024-10-01 20:23:17.055454] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:24:22.017 [2024-10-01 20:23:17.055573] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:24:22.017 [2024-10-01 20:23:17.055727] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:24:22.017 [2024-10-01 20:23:17.055751] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:24:22.017 [2024-10-01 20:23:17.055770] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:22.017 "name": "Existed_Raid", 00:24:22.017 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:22.017 "strip_size_kb": 64, 00:24:22.017 "state": "configuring", 00:24:22.017 "raid_level": "concat", 00:24:22.017 "superblock": false, 00:24:22.017 "num_base_bdevs": 4, 00:24:22.017 "num_base_bdevs_discovered": 1, 00:24:22.017 "num_base_bdevs_operational": 4, 00:24:22.017 "base_bdevs_list": [ 00:24:22.017 { 00:24:22.017 "name": "BaseBdev1", 00:24:22.017 "uuid": "7684bd78-5c0f-430d-89a6-7bbd2b151bbd", 00:24:22.017 "is_configured": true, 00:24:22.017 "data_offset": 0, 00:24:22.017 "data_size": 65536 00:24:22.017 }, 00:24:22.017 { 00:24:22.017 "name": "BaseBdev2", 00:24:22.017 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:22.017 "is_configured": false, 00:24:22.017 "data_offset": 0, 00:24:22.017 "data_size": 0 00:24:22.017 }, 00:24:22.017 { 00:24:22.017 "name": "BaseBdev3", 00:24:22.017 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:22.017 "is_configured": false, 00:24:22.017 "data_offset": 0, 00:24:22.017 "data_size": 0 00:24:22.017 }, 00:24:22.017 { 00:24:22.017 "name": "BaseBdev4", 00:24:22.017 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:22.017 "is_configured": false, 00:24:22.017 "data_offset": 0, 00:24:22.017 "data_size": 0 00:24:22.017 } 00:24:22.017 ] 00:24:22.017 }' 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:22.017 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:22.586 [2024-10-01 20:23:17.598648] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:24:22.586 BaseBdev2 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:22.586 [ 00:24:22.586 { 00:24:22.586 "name": "BaseBdev2", 00:24:22.586 "aliases": [ 00:24:22.586 "a3d1bdd1-30a9-462f-8b1e-18dd013ba067" 00:24:22.586 ], 00:24:22.586 "product_name": "Malloc disk", 00:24:22.586 "block_size": 512, 00:24:22.586 "num_blocks": 65536, 00:24:22.586 "uuid": "a3d1bdd1-30a9-462f-8b1e-18dd013ba067", 00:24:22.586 "assigned_rate_limits": { 00:24:22.586 "rw_ios_per_sec": 0, 00:24:22.586 "rw_mbytes_per_sec": 0, 00:24:22.586 "r_mbytes_per_sec": 0, 00:24:22.586 "w_mbytes_per_sec": 0 00:24:22.586 }, 00:24:22.586 "claimed": true, 00:24:22.586 "claim_type": "exclusive_write", 00:24:22.586 "zoned": false, 00:24:22.586 "supported_io_types": { 00:24:22.586 "read": true, 00:24:22.586 "write": true, 00:24:22.586 "unmap": true, 00:24:22.586 "flush": true, 00:24:22.586 "reset": true, 00:24:22.586 "nvme_admin": false, 00:24:22.586 "nvme_io": false, 00:24:22.586 "nvme_io_md": false, 00:24:22.586 "write_zeroes": true, 00:24:22.586 "zcopy": true, 00:24:22.586 "get_zone_info": false, 00:24:22.586 "zone_management": false, 00:24:22.586 "zone_append": false, 00:24:22.586 "compare": false, 00:24:22.586 "compare_and_write": false, 00:24:22.586 "abort": true, 00:24:22.586 "seek_hole": false, 00:24:22.586 "seek_data": false, 00:24:22.586 "copy": true, 00:24:22.586 "nvme_iov_md": false 00:24:22.586 }, 00:24:22.586 "memory_domains": [ 00:24:22.586 { 00:24:22.586 "dma_device_id": "system", 00:24:22.586 "dma_device_type": 1 00:24:22.586 }, 00:24:22.586 { 00:24:22.586 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:22.586 "dma_device_type": 2 00:24:22.586 } 00:24:22.586 ], 00:24:22.586 "driver_specific": {} 00:24:22.586 } 00:24:22.586 ] 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:22.586 "name": "Existed_Raid", 00:24:22.586 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:22.586 "strip_size_kb": 64, 00:24:22.586 "state": "configuring", 00:24:22.586 "raid_level": "concat", 00:24:22.586 "superblock": false, 00:24:22.586 "num_base_bdevs": 4, 00:24:22.586 "num_base_bdevs_discovered": 2, 00:24:22.586 "num_base_bdevs_operational": 4, 00:24:22.586 "base_bdevs_list": [ 00:24:22.586 { 00:24:22.586 "name": "BaseBdev1", 00:24:22.586 "uuid": "7684bd78-5c0f-430d-89a6-7bbd2b151bbd", 00:24:22.586 "is_configured": true, 00:24:22.586 "data_offset": 0, 00:24:22.586 "data_size": 65536 00:24:22.586 }, 00:24:22.586 { 00:24:22.586 "name": "BaseBdev2", 00:24:22.586 "uuid": "a3d1bdd1-30a9-462f-8b1e-18dd013ba067", 00:24:22.586 "is_configured": true, 00:24:22.586 "data_offset": 0, 00:24:22.586 "data_size": 65536 00:24:22.586 }, 00:24:22.586 { 00:24:22.586 "name": "BaseBdev3", 00:24:22.586 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:22.586 "is_configured": false, 00:24:22.586 "data_offset": 0, 00:24:22.586 "data_size": 0 00:24:22.586 }, 00:24:22.586 { 00:24:22.586 "name": "BaseBdev4", 00:24:22.586 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:22.586 "is_configured": false, 00:24:22.586 "data_offset": 0, 00:24:22.586 "data_size": 0 00:24:22.586 } 00:24:22.586 ] 00:24:22.586 }' 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:22.586 20:23:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:23.154 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:24:23.154 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:23.154 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:23.154 [2024-10-01 20:23:18.182746] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:24:23.154 BaseBdev3 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:23.155 [ 00:24:23.155 { 00:24:23.155 "name": "BaseBdev3", 00:24:23.155 "aliases": [ 00:24:23.155 "cb65c439-a6fe-428c-b4b9-2ca2bbd73eab" 00:24:23.155 ], 00:24:23.155 "product_name": "Malloc disk", 00:24:23.155 "block_size": 512, 00:24:23.155 "num_blocks": 65536, 00:24:23.155 "uuid": "cb65c439-a6fe-428c-b4b9-2ca2bbd73eab", 00:24:23.155 "assigned_rate_limits": { 00:24:23.155 "rw_ios_per_sec": 0, 00:24:23.155 "rw_mbytes_per_sec": 0, 00:24:23.155 "r_mbytes_per_sec": 0, 00:24:23.155 "w_mbytes_per_sec": 0 00:24:23.155 }, 00:24:23.155 "claimed": true, 00:24:23.155 "claim_type": "exclusive_write", 00:24:23.155 "zoned": false, 00:24:23.155 "supported_io_types": { 00:24:23.155 "read": true, 00:24:23.155 "write": true, 00:24:23.155 "unmap": true, 00:24:23.155 "flush": true, 00:24:23.155 "reset": true, 00:24:23.155 "nvme_admin": false, 00:24:23.155 "nvme_io": false, 00:24:23.155 "nvme_io_md": false, 00:24:23.155 "write_zeroes": true, 00:24:23.155 "zcopy": true, 00:24:23.155 "get_zone_info": false, 00:24:23.155 "zone_management": false, 00:24:23.155 "zone_append": false, 00:24:23.155 "compare": false, 00:24:23.155 "compare_and_write": false, 00:24:23.155 "abort": true, 00:24:23.155 "seek_hole": false, 00:24:23.155 "seek_data": false, 00:24:23.155 "copy": true, 00:24:23.155 "nvme_iov_md": false 00:24:23.155 }, 00:24:23.155 "memory_domains": [ 00:24:23.155 { 00:24:23.155 "dma_device_id": "system", 00:24:23.155 "dma_device_type": 1 00:24:23.155 }, 00:24:23.155 { 00:24:23.155 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:23.155 "dma_device_type": 2 00:24:23.155 } 00:24:23.155 ], 00:24:23.155 "driver_specific": {} 00:24:23.155 } 00:24:23.155 ] 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:23.155 "name": "Existed_Raid", 00:24:23.155 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:23.155 "strip_size_kb": 64, 00:24:23.155 "state": "configuring", 00:24:23.155 "raid_level": "concat", 00:24:23.155 "superblock": false, 00:24:23.155 "num_base_bdevs": 4, 00:24:23.155 "num_base_bdevs_discovered": 3, 00:24:23.155 "num_base_bdevs_operational": 4, 00:24:23.155 "base_bdevs_list": [ 00:24:23.155 { 00:24:23.155 "name": "BaseBdev1", 00:24:23.155 "uuid": "7684bd78-5c0f-430d-89a6-7bbd2b151bbd", 00:24:23.155 "is_configured": true, 00:24:23.155 "data_offset": 0, 00:24:23.155 "data_size": 65536 00:24:23.155 }, 00:24:23.155 { 00:24:23.155 "name": "BaseBdev2", 00:24:23.155 "uuid": "a3d1bdd1-30a9-462f-8b1e-18dd013ba067", 00:24:23.155 "is_configured": true, 00:24:23.155 "data_offset": 0, 00:24:23.155 "data_size": 65536 00:24:23.155 }, 00:24:23.155 { 00:24:23.155 "name": "BaseBdev3", 00:24:23.155 "uuid": "cb65c439-a6fe-428c-b4b9-2ca2bbd73eab", 00:24:23.155 "is_configured": true, 00:24:23.155 "data_offset": 0, 00:24:23.155 "data_size": 65536 00:24:23.155 }, 00:24:23.155 { 00:24:23.155 "name": "BaseBdev4", 00:24:23.155 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:23.155 "is_configured": false, 00:24:23.155 "data_offset": 0, 00:24:23.155 "data_size": 0 00:24:23.155 } 00:24:23.155 ] 00:24:23.155 }' 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:23.155 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:23.722 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:24:23.722 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:23.723 [2024-10-01 20:23:18.770179] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:24:23.723 [2024-10-01 20:23:18.770287] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:24:23.723 [2024-10-01 20:23:18.770301] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:24:23.723 [2024-10-01 20:23:18.770647] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:24:23.723 [2024-10-01 20:23:18.770948] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:24:23.723 [2024-10-01 20:23:18.770970] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:24:23.723 [2024-10-01 20:23:18.771327] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:23.723 BaseBdev4 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:23.723 [ 00:24:23.723 { 00:24:23.723 "name": "BaseBdev4", 00:24:23.723 "aliases": [ 00:24:23.723 "8d0a1547-4921-4542-a163-d7168978484a" 00:24:23.723 ], 00:24:23.723 "product_name": "Malloc disk", 00:24:23.723 "block_size": 512, 00:24:23.723 "num_blocks": 65536, 00:24:23.723 "uuid": "8d0a1547-4921-4542-a163-d7168978484a", 00:24:23.723 "assigned_rate_limits": { 00:24:23.723 "rw_ios_per_sec": 0, 00:24:23.723 "rw_mbytes_per_sec": 0, 00:24:23.723 "r_mbytes_per_sec": 0, 00:24:23.723 "w_mbytes_per_sec": 0 00:24:23.723 }, 00:24:23.723 "claimed": true, 00:24:23.723 "claim_type": "exclusive_write", 00:24:23.723 "zoned": false, 00:24:23.723 "supported_io_types": { 00:24:23.723 "read": true, 00:24:23.723 "write": true, 00:24:23.723 "unmap": true, 00:24:23.723 "flush": true, 00:24:23.723 "reset": true, 00:24:23.723 "nvme_admin": false, 00:24:23.723 "nvme_io": false, 00:24:23.723 "nvme_io_md": false, 00:24:23.723 "write_zeroes": true, 00:24:23.723 "zcopy": true, 00:24:23.723 "get_zone_info": false, 00:24:23.723 "zone_management": false, 00:24:23.723 "zone_append": false, 00:24:23.723 "compare": false, 00:24:23.723 "compare_and_write": false, 00:24:23.723 "abort": true, 00:24:23.723 "seek_hole": false, 00:24:23.723 "seek_data": false, 00:24:23.723 "copy": true, 00:24:23.723 "nvme_iov_md": false 00:24:23.723 }, 00:24:23.723 "memory_domains": [ 00:24:23.723 { 00:24:23.723 "dma_device_id": "system", 00:24:23.723 "dma_device_type": 1 00:24:23.723 }, 00:24:23.723 { 00:24:23.723 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:23.723 "dma_device_type": 2 00:24:23.723 } 00:24:23.723 ], 00:24:23.723 "driver_specific": {} 00:24:23.723 } 00:24:23.723 ] 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:23.723 "name": "Existed_Raid", 00:24:23.723 "uuid": "b31203bf-5be5-4a03-95a5-e47fd5c7896f", 00:24:23.723 "strip_size_kb": 64, 00:24:23.723 "state": "online", 00:24:23.723 "raid_level": "concat", 00:24:23.723 "superblock": false, 00:24:23.723 "num_base_bdevs": 4, 00:24:23.723 "num_base_bdevs_discovered": 4, 00:24:23.723 "num_base_bdevs_operational": 4, 00:24:23.723 "base_bdevs_list": [ 00:24:23.723 { 00:24:23.723 "name": "BaseBdev1", 00:24:23.723 "uuid": "7684bd78-5c0f-430d-89a6-7bbd2b151bbd", 00:24:23.723 "is_configured": true, 00:24:23.723 "data_offset": 0, 00:24:23.723 "data_size": 65536 00:24:23.723 }, 00:24:23.723 { 00:24:23.723 "name": "BaseBdev2", 00:24:23.723 "uuid": "a3d1bdd1-30a9-462f-8b1e-18dd013ba067", 00:24:23.723 "is_configured": true, 00:24:23.723 "data_offset": 0, 00:24:23.723 "data_size": 65536 00:24:23.723 }, 00:24:23.723 { 00:24:23.723 "name": "BaseBdev3", 00:24:23.723 "uuid": "cb65c439-a6fe-428c-b4b9-2ca2bbd73eab", 00:24:23.723 "is_configured": true, 00:24:23.723 "data_offset": 0, 00:24:23.723 "data_size": 65536 00:24:23.723 }, 00:24:23.723 { 00:24:23.723 "name": "BaseBdev4", 00:24:23.723 "uuid": "8d0a1547-4921-4542-a163-d7168978484a", 00:24:23.723 "is_configured": true, 00:24:23.723 "data_offset": 0, 00:24:23.723 "data_size": 65536 00:24:23.723 } 00:24:23.723 ] 00:24:23.723 }' 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:23.723 20:23:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:24.289 [2024-10-01 20:23:19.342913] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:24.289 "name": "Existed_Raid", 00:24:24.289 "aliases": [ 00:24:24.289 "b31203bf-5be5-4a03-95a5-e47fd5c7896f" 00:24:24.289 ], 00:24:24.289 "product_name": "Raid Volume", 00:24:24.289 "block_size": 512, 00:24:24.289 "num_blocks": 262144, 00:24:24.289 "uuid": "b31203bf-5be5-4a03-95a5-e47fd5c7896f", 00:24:24.289 "assigned_rate_limits": { 00:24:24.289 "rw_ios_per_sec": 0, 00:24:24.289 "rw_mbytes_per_sec": 0, 00:24:24.289 "r_mbytes_per_sec": 0, 00:24:24.289 "w_mbytes_per_sec": 0 00:24:24.289 }, 00:24:24.289 "claimed": false, 00:24:24.289 "zoned": false, 00:24:24.289 "supported_io_types": { 00:24:24.289 "read": true, 00:24:24.289 "write": true, 00:24:24.289 "unmap": true, 00:24:24.289 "flush": true, 00:24:24.289 "reset": true, 00:24:24.289 "nvme_admin": false, 00:24:24.289 "nvme_io": false, 00:24:24.289 "nvme_io_md": false, 00:24:24.289 "write_zeroes": true, 00:24:24.289 "zcopy": false, 00:24:24.289 "get_zone_info": false, 00:24:24.289 "zone_management": false, 00:24:24.289 "zone_append": false, 00:24:24.289 "compare": false, 00:24:24.289 "compare_and_write": false, 00:24:24.289 "abort": false, 00:24:24.289 "seek_hole": false, 00:24:24.289 "seek_data": false, 00:24:24.289 "copy": false, 00:24:24.289 "nvme_iov_md": false 00:24:24.289 }, 00:24:24.289 "memory_domains": [ 00:24:24.289 { 00:24:24.289 "dma_device_id": "system", 00:24:24.289 "dma_device_type": 1 00:24:24.289 }, 00:24:24.289 { 00:24:24.289 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:24.289 "dma_device_type": 2 00:24:24.289 }, 00:24:24.289 { 00:24:24.289 "dma_device_id": "system", 00:24:24.289 "dma_device_type": 1 00:24:24.289 }, 00:24:24.289 { 00:24:24.289 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:24.289 "dma_device_type": 2 00:24:24.289 }, 00:24:24.289 { 00:24:24.289 "dma_device_id": "system", 00:24:24.289 "dma_device_type": 1 00:24:24.289 }, 00:24:24.289 { 00:24:24.289 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:24.289 "dma_device_type": 2 00:24:24.289 }, 00:24:24.289 { 00:24:24.289 "dma_device_id": "system", 00:24:24.289 "dma_device_type": 1 00:24:24.289 }, 00:24:24.289 { 00:24:24.289 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:24.289 "dma_device_type": 2 00:24:24.289 } 00:24:24.289 ], 00:24:24.289 "driver_specific": { 00:24:24.289 "raid": { 00:24:24.289 "uuid": "b31203bf-5be5-4a03-95a5-e47fd5c7896f", 00:24:24.289 "strip_size_kb": 64, 00:24:24.289 "state": "online", 00:24:24.289 "raid_level": "concat", 00:24:24.289 "superblock": false, 00:24:24.289 "num_base_bdevs": 4, 00:24:24.289 "num_base_bdevs_discovered": 4, 00:24:24.289 "num_base_bdevs_operational": 4, 00:24:24.289 "base_bdevs_list": [ 00:24:24.289 { 00:24:24.289 "name": "BaseBdev1", 00:24:24.289 "uuid": "7684bd78-5c0f-430d-89a6-7bbd2b151bbd", 00:24:24.289 "is_configured": true, 00:24:24.289 "data_offset": 0, 00:24:24.289 "data_size": 65536 00:24:24.289 }, 00:24:24.289 { 00:24:24.289 "name": "BaseBdev2", 00:24:24.289 "uuid": "a3d1bdd1-30a9-462f-8b1e-18dd013ba067", 00:24:24.289 "is_configured": true, 00:24:24.289 "data_offset": 0, 00:24:24.289 "data_size": 65536 00:24:24.289 }, 00:24:24.289 { 00:24:24.289 "name": "BaseBdev3", 00:24:24.289 "uuid": "cb65c439-a6fe-428c-b4b9-2ca2bbd73eab", 00:24:24.289 "is_configured": true, 00:24:24.289 "data_offset": 0, 00:24:24.289 "data_size": 65536 00:24:24.289 }, 00:24:24.289 { 00:24:24.289 "name": "BaseBdev4", 00:24:24.289 "uuid": "8d0a1547-4921-4542-a163-d7168978484a", 00:24:24.289 "is_configured": true, 00:24:24.289 "data_offset": 0, 00:24:24.289 "data_size": 65536 00:24:24.289 } 00:24:24.289 ] 00:24:24.289 } 00:24:24.289 } 00:24:24.289 }' 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:24:24.289 BaseBdev2 00:24:24.289 BaseBdev3 00:24:24.289 BaseBdev4' 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:24.289 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:24.546 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:24.546 [2024-10-01 20:23:19.718642] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:24:24.546 [2024-10-01 20:23:19.719834] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:24:24.546 [2024-10-01 20:23:19.719936] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:24.804 "name": "Existed_Raid", 00:24:24.804 "uuid": "b31203bf-5be5-4a03-95a5-e47fd5c7896f", 00:24:24.804 "strip_size_kb": 64, 00:24:24.804 "state": "offline", 00:24:24.804 "raid_level": "concat", 00:24:24.804 "superblock": false, 00:24:24.804 "num_base_bdevs": 4, 00:24:24.804 "num_base_bdevs_discovered": 3, 00:24:24.804 "num_base_bdevs_operational": 3, 00:24:24.804 "base_bdevs_list": [ 00:24:24.804 { 00:24:24.804 "name": null, 00:24:24.804 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:24.804 "is_configured": false, 00:24:24.804 "data_offset": 0, 00:24:24.804 "data_size": 65536 00:24:24.804 }, 00:24:24.804 { 00:24:24.804 "name": "BaseBdev2", 00:24:24.804 "uuid": "a3d1bdd1-30a9-462f-8b1e-18dd013ba067", 00:24:24.804 "is_configured": true, 00:24:24.804 "data_offset": 0, 00:24:24.804 "data_size": 65536 00:24:24.804 }, 00:24:24.804 { 00:24:24.804 "name": "BaseBdev3", 00:24:24.804 "uuid": "cb65c439-a6fe-428c-b4b9-2ca2bbd73eab", 00:24:24.804 "is_configured": true, 00:24:24.804 "data_offset": 0, 00:24:24.804 "data_size": 65536 00:24:24.804 }, 00:24:24.804 { 00:24:24.804 "name": "BaseBdev4", 00:24:24.804 "uuid": "8d0a1547-4921-4542-a163-d7168978484a", 00:24:24.804 "is_configured": true, 00:24:24.804 "data_offset": 0, 00:24:24.804 "data_size": 65536 00:24:24.804 } 00:24:24.804 ] 00:24:24.804 }' 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:24.804 20:23:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.369 [2024-10-01 20:23:20.462603] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:25.369 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.369 [2024-10-01 20:23:20.617133] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:24:25.627 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:25.627 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:24:25.627 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:24:25.627 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:24:25.627 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:25.627 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:25.627 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.627 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:25.627 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:24:25.627 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:24:25.628 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:24:25.628 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:25.628 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.628 [2024-10-01 20:23:20.766143] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:24:25.628 [2024-10-01 20:23:20.766330] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:24:25.628 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:25.628 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:24:25.628 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:24:25.628 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:25.628 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:24:25.628 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:25.628 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.628 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.887 BaseBdev2 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.887 [ 00:24:25.887 { 00:24:25.887 "name": "BaseBdev2", 00:24:25.887 "aliases": [ 00:24:25.887 "138795bb-0cdb-4865-8dac-f5c15361950f" 00:24:25.887 ], 00:24:25.887 "product_name": "Malloc disk", 00:24:25.887 "block_size": 512, 00:24:25.887 "num_blocks": 65536, 00:24:25.887 "uuid": "138795bb-0cdb-4865-8dac-f5c15361950f", 00:24:25.887 "assigned_rate_limits": { 00:24:25.887 "rw_ios_per_sec": 0, 00:24:25.887 "rw_mbytes_per_sec": 0, 00:24:25.887 "r_mbytes_per_sec": 0, 00:24:25.887 "w_mbytes_per_sec": 0 00:24:25.887 }, 00:24:25.887 "claimed": false, 00:24:25.887 "zoned": false, 00:24:25.887 "supported_io_types": { 00:24:25.887 "read": true, 00:24:25.887 "write": true, 00:24:25.887 "unmap": true, 00:24:25.887 "flush": true, 00:24:25.887 "reset": true, 00:24:25.887 "nvme_admin": false, 00:24:25.887 "nvme_io": false, 00:24:25.887 "nvme_io_md": false, 00:24:25.887 "write_zeroes": true, 00:24:25.887 "zcopy": true, 00:24:25.887 "get_zone_info": false, 00:24:25.887 "zone_management": false, 00:24:25.887 "zone_append": false, 00:24:25.887 "compare": false, 00:24:25.887 "compare_and_write": false, 00:24:25.887 "abort": true, 00:24:25.887 "seek_hole": false, 00:24:25.887 "seek_data": false, 00:24:25.887 "copy": true, 00:24:25.887 "nvme_iov_md": false 00:24:25.887 }, 00:24:25.887 "memory_domains": [ 00:24:25.887 { 00:24:25.887 "dma_device_id": "system", 00:24:25.887 "dma_device_type": 1 00:24:25.887 }, 00:24:25.887 { 00:24:25.887 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:25.887 "dma_device_type": 2 00:24:25.887 } 00:24:25.887 ], 00:24:25.887 "driver_specific": {} 00:24:25.887 } 00:24:25.887 ] 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:25.887 20:23:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.887 BaseBdev3 00:24:25.887 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:25.887 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:24:25.887 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:24:25.887 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:25.887 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:24:25.887 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:25.887 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:25.887 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:25.887 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:25.887 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.887 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:25.887 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:24:25.887 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:25.887 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.887 [ 00:24:25.887 { 00:24:25.887 "name": "BaseBdev3", 00:24:25.887 "aliases": [ 00:24:25.887 "03b4c4db-4b3c-4fbf-a828-636d274d1e90" 00:24:25.887 ], 00:24:25.887 "product_name": "Malloc disk", 00:24:25.887 "block_size": 512, 00:24:25.887 "num_blocks": 65536, 00:24:25.887 "uuid": "03b4c4db-4b3c-4fbf-a828-636d274d1e90", 00:24:25.888 "assigned_rate_limits": { 00:24:25.888 "rw_ios_per_sec": 0, 00:24:25.888 "rw_mbytes_per_sec": 0, 00:24:25.888 "r_mbytes_per_sec": 0, 00:24:25.888 "w_mbytes_per_sec": 0 00:24:25.888 }, 00:24:25.888 "claimed": false, 00:24:25.888 "zoned": false, 00:24:25.888 "supported_io_types": { 00:24:25.888 "read": true, 00:24:25.888 "write": true, 00:24:25.888 "unmap": true, 00:24:25.888 "flush": true, 00:24:25.888 "reset": true, 00:24:25.888 "nvme_admin": false, 00:24:25.888 "nvme_io": false, 00:24:25.888 "nvme_io_md": false, 00:24:25.888 "write_zeroes": true, 00:24:25.888 "zcopy": true, 00:24:25.888 "get_zone_info": false, 00:24:25.888 "zone_management": false, 00:24:25.888 "zone_append": false, 00:24:25.888 "compare": false, 00:24:25.888 "compare_and_write": false, 00:24:25.888 "abort": true, 00:24:25.888 "seek_hole": false, 00:24:25.888 "seek_data": false, 00:24:25.888 "copy": true, 00:24:25.888 "nvme_iov_md": false 00:24:25.888 }, 00:24:25.888 "memory_domains": [ 00:24:25.888 { 00:24:25.888 "dma_device_id": "system", 00:24:25.888 "dma_device_type": 1 00:24:25.888 }, 00:24:25.888 { 00:24:25.888 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:25.888 "dma_device_type": 2 00:24:25.888 } 00:24:25.888 ], 00:24:25.888 "driver_specific": {} 00:24:25.888 } 00:24:25.888 ] 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.888 BaseBdev4 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:25.888 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:25.888 [ 00:24:25.888 { 00:24:25.888 "name": "BaseBdev4", 00:24:25.888 "aliases": [ 00:24:25.888 "4931312f-2000-4f27-b4a3-6a3890219c86" 00:24:25.888 ], 00:24:25.888 "product_name": "Malloc disk", 00:24:25.888 "block_size": 512, 00:24:25.888 "num_blocks": 65536, 00:24:25.888 "uuid": "4931312f-2000-4f27-b4a3-6a3890219c86", 00:24:25.888 "assigned_rate_limits": { 00:24:25.888 "rw_ios_per_sec": 0, 00:24:25.888 "rw_mbytes_per_sec": 0, 00:24:25.888 "r_mbytes_per_sec": 0, 00:24:25.888 "w_mbytes_per_sec": 0 00:24:25.888 }, 00:24:25.888 "claimed": false, 00:24:25.888 "zoned": false, 00:24:25.888 "supported_io_types": { 00:24:25.888 "read": true, 00:24:25.888 "write": true, 00:24:25.888 "unmap": true, 00:24:25.888 "flush": true, 00:24:25.888 "reset": true, 00:24:25.888 "nvme_admin": false, 00:24:25.888 "nvme_io": false, 00:24:25.888 "nvme_io_md": false, 00:24:25.888 "write_zeroes": true, 00:24:25.888 "zcopy": true, 00:24:25.888 "get_zone_info": false, 00:24:25.888 "zone_management": false, 00:24:25.888 "zone_append": false, 00:24:25.888 "compare": false, 00:24:25.888 "compare_and_write": false, 00:24:25.888 "abort": true, 00:24:25.888 "seek_hole": false, 00:24:25.888 "seek_data": false, 00:24:25.888 "copy": true, 00:24:25.888 "nvme_iov_md": false 00:24:25.888 }, 00:24:25.888 "memory_domains": [ 00:24:25.888 { 00:24:25.888 "dma_device_id": "system", 00:24:25.888 "dma_device_type": 1 00:24:25.888 }, 00:24:25.888 { 00:24:25.888 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:25.888 "dma_device_type": 2 00:24:25.888 } 00:24:25.888 ], 00:24:26.147 "driver_specific": {} 00:24:26.147 } 00:24:26.147 ] 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:26.147 [2024-10-01 20:23:21.145869] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:24:26.147 [2024-10-01 20:23:21.146052] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:24:26.147 [2024-10-01 20:23:21.146186] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:24:26.147 [2024-10-01 20:23:21.148785] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:24:26.147 [2024-10-01 20:23:21.148977] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:26.147 "name": "Existed_Raid", 00:24:26.147 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:26.147 "strip_size_kb": 64, 00:24:26.147 "state": "configuring", 00:24:26.147 "raid_level": "concat", 00:24:26.147 "superblock": false, 00:24:26.147 "num_base_bdevs": 4, 00:24:26.147 "num_base_bdevs_discovered": 3, 00:24:26.147 "num_base_bdevs_operational": 4, 00:24:26.147 "base_bdevs_list": [ 00:24:26.147 { 00:24:26.147 "name": "BaseBdev1", 00:24:26.147 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:26.147 "is_configured": false, 00:24:26.147 "data_offset": 0, 00:24:26.147 "data_size": 0 00:24:26.147 }, 00:24:26.147 { 00:24:26.147 "name": "BaseBdev2", 00:24:26.147 "uuid": "138795bb-0cdb-4865-8dac-f5c15361950f", 00:24:26.147 "is_configured": true, 00:24:26.147 "data_offset": 0, 00:24:26.147 "data_size": 65536 00:24:26.147 }, 00:24:26.147 { 00:24:26.147 "name": "BaseBdev3", 00:24:26.147 "uuid": "03b4c4db-4b3c-4fbf-a828-636d274d1e90", 00:24:26.147 "is_configured": true, 00:24:26.147 "data_offset": 0, 00:24:26.147 "data_size": 65536 00:24:26.147 }, 00:24:26.147 { 00:24:26.147 "name": "BaseBdev4", 00:24:26.147 "uuid": "4931312f-2000-4f27-b4a3-6a3890219c86", 00:24:26.147 "is_configured": true, 00:24:26.147 "data_offset": 0, 00:24:26.147 "data_size": 65536 00:24:26.147 } 00:24:26.147 ] 00:24:26.147 }' 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:26.147 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:26.713 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:24:26.713 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:26.713 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:26.713 [2024-10-01 20:23:21.698077] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:24:26.713 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:26.713 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:26.713 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:26.713 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:26.713 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:26.713 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:26.713 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:26.713 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:26.713 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:26.713 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:26.713 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:26.713 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:26.714 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:26.714 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:26.714 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:26.714 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:26.714 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:26.714 "name": "Existed_Raid", 00:24:26.714 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:26.714 "strip_size_kb": 64, 00:24:26.714 "state": "configuring", 00:24:26.714 "raid_level": "concat", 00:24:26.714 "superblock": false, 00:24:26.714 "num_base_bdevs": 4, 00:24:26.714 "num_base_bdevs_discovered": 2, 00:24:26.714 "num_base_bdevs_operational": 4, 00:24:26.714 "base_bdevs_list": [ 00:24:26.714 { 00:24:26.714 "name": "BaseBdev1", 00:24:26.714 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:26.714 "is_configured": false, 00:24:26.714 "data_offset": 0, 00:24:26.714 "data_size": 0 00:24:26.714 }, 00:24:26.714 { 00:24:26.714 "name": null, 00:24:26.714 "uuid": "138795bb-0cdb-4865-8dac-f5c15361950f", 00:24:26.714 "is_configured": false, 00:24:26.714 "data_offset": 0, 00:24:26.714 "data_size": 65536 00:24:26.714 }, 00:24:26.714 { 00:24:26.714 "name": "BaseBdev3", 00:24:26.714 "uuid": "03b4c4db-4b3c-4fbf-a828-636d274d1e90", 00:24:26.714 "is_configured": true, 00:24:26.714 "data_offset": 0, 00:24:26.714 "data_size": 65536 00:24:26.714 }, 00:24:26.714 { 00:24:26.714 "name": "BaseBdev4", 00:24:26.714 "uuid": "4931312f-2000-4f27-b4a3-6a3890219c86", 00:24:26.714 "is_configured": true, 00:24:26.714 "data_offset": 0, 00:24:26.714 "data_size": 65536 00:24:26.714 } 00:24:26.714 ] 00:24:26.714 }' 00:24:26.714 20:23:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:26.714 20:23:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:26.996 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:26.996 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:24:26.996 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:26.996 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:26.996 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:27.273 [2024-10-01 20:23:22.296821] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:24:27.273 BaseBdev1 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:27.273 [ 00:24:27.273 { 00:24:27.273 "name": "BaseBdev1", 00:24:27.273 "aliases": [ 00:24:27.273 "9945eea1-903a-4100-bcdc-514dbb8ba27e" 00:24:27.273 ], 00:24:27.273 "product_name": "Malloc disk", 00:24:27.273 "block_size": 512, 00:24:27.273 "num_blocks": 65536, 00:24:27.273 "uuid": "9945eea1-903a-4100-bcdc-514dbb8ba27e", 00:24:27.273 "assigned_rate_limits": { 00:24:27.273 "rw_ios_per_sec": 0, 00:24:27.273 "rw_mbytes_per_sec": 0, 00:24:27.273 "r_mbytes_per_sec": 0, 00:24:27.273 "w_mbytes_per_sec": 0 00:24:27.273 }, 00:24:27.273 "claimed": true, 00:24:27.273 "claim_type": "exclusive_write", 00:24:27.273 "zoned": false, 00:24:27.273 "supported_io_types": { 00:24:27.273 "read": true, 00:24:27.273 "write": true, 00:24:27.273 "unmap": true, 00:24:27.273 "flush": true, 00:24:27.273 "reset": true, 00:24:27.273 "nvme_admin": false, 00:24:27.273 "nvme_io": false, 00:24:27.273 "nvme_io_md": false, 00:24:27.273 "write_zeroes": true, 00:24:27.273 "zcopy": true, 00:24:27.273 "get_zone_info": false, 00:24:27.273 "zone_management": false, 00:24:27.273 "zone_append": false, 00:24:27.273 "compare": false, 00:24:27.273 "compare_and_write": false, 00:24:27.273 "abort": true, 00:24:27.273 "seek_hole": false, 00:24:27.273 "seek_data": false, 00:24:27.273 "copy": true, 00:24:27.273 "nvme_iov_md": false 00:24:27.273 }, 00:24:27.273 "memory_domains": [ 00:24:27.273 { 00:24:27.273 "dma_device_id": "system", 00:24:27.273 "dma_device_type": 1 00:24:27.273 }, 00:24:27.273 { 00:24:27.273 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:27.273 "dma_device_type": 2 00:24:27.273 } 00:24:27.273 ], 00:24:27.273 "driver_specific": {} 00:24:27.273 } 00:24:27.273 ] 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:27.273 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:27.274 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:27.274 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:27.274 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:27.274 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:27.274 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:27.274 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:27.274 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:27.274 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:27.274 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:27.274 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:27.274 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:27.274 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:27.274 "name": "Existed_Raid", 00:24:27.274 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:27.274 "strip_size_kb": 64, 00:24:27.274 "state": "configuring", 00:24:27.274 "raid_level": "concat", 00:24:27.274 "superblock": false, 00:24:27.274 "num_base_bdevs": 4, 00:24:27.274 "num_base_bdevs_discovered": 3, 00:24:27.274 "num_base_bdevs_operational": 4, 00:24:27.274 "base_bdevs_list": [ 00:24:27.274 { 00:24:27.274 "name": "BaseBdev1", 00:24:27.274 "uuid": "9945eea1-903a-4100-bcdc-514dbb8ba27e", 00:24:27.274 "is_configured": true, 00:24:27.274 "data_offset": 0, 00:24:27.274 "data_size": 65536 00:24:27.274 }, 00:24:27.274 { 00:24:27.274 "name": null, 00:24:27.274 "uuid": "138795bb-0cdb-4865-8dac-f5c15361950f", 00:24:27.274 "is_configured": false, 00:24:27.274 "data_offset": 0, 00:24:27.274 "data_size": 65536 00:24:27.274 }, 00:24:27.274 { 00:24:27.274 "name": "BaseBdev3", 00:24:27.274 "uuid": "03b4c4db-4b3c-4fbf-a828-636d274d1e90", 00:24:27.274 "is_configured": true, 00:24:27.274 "data_offset": 0, 00:24:27.274 "data_size": 65536 00:24:27.274 }, 00:24:27.274 { 00:24:27.274 "name": "BaseBdev4", 00:24:27.274 "uuid": "4931312f-2000-4f27-b4a3-6a3890219c86", 00:24:27.274 "is_configured": true, 00:24:27.274 "data_offset": 0, 00:24:27.274 "data_size": 65536 00:24:27.274 } 00:24:27.274 ] 00:24:27.274 }' 00:24:27.274 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:27.274 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:27.840 [2024-10-01 20:23:22.917129] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:27.840 "name": "Existed_Raid", 00:24:27.840 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:27.840 "strip_size_kb": 64, 00:24:27.840 "state": "configuring", 00:24:27.840 "raid_level": "concat", 00:24:27.840 "superblock": false, 00:24:27.840 "num_base_bdevs": 4, 00:24:27.840 "num_base_bdevs_discovered": 2, 00:24:27.840 "num_base_bdevs_operational": 4, 00:24:27.840 "base_bdevs_list": [ 00:24:27.840 { 00:24:27.840 "name": "BaseBdev1", 00:24:27.840 "uuid": "9945eea1-903a-4100-bcdc-514dbb8ba27e", 00:24:27.840 "is_configured": true, 00:24:27.840 "data_offset": 0, 00:24:27.840 "data_size": 65536 00:24:27.840 }, 00:24:27.840 { 00:24:27.840 "name": null, 00:24:27.840 "uuid": "138795bb-0cdb-4865-8dac-f5c15361950f", 00:24:27.840 "is_configured": false, 00:24:27.840 "data_offset": 0, 00:24:27.840 "data_size": 65536 00:24:27.840 }, 00:24:27.840 { 00:24:27.840 "name": null, 00:24:27.840 "uuid": "03b4c4db-4b3c-4fbf-a828-636d274d1e90", 00:24:27.840 "is_configured": false, 00:24:27.840 "data_offset": 0, 00:24:27.840 "data_size": 65536 00:24:27.840 }, 00:24:27.840 { 00:24:27.840 "name": "BaseBdev4", 00:24:27.840 "uuid": "4931312f-2000-4f27-b4a3-6a3890219c86", 00:24:27.840 "is_configured": true, 00:24:27.840 "data_offset": 0, 00:24:27.840 "data_size": 65536 00:24:27.840 } 00:24:27.840 ] 00:24:27.840 }' 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:27.840 20:23:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:28.407 [2024-10-01 20:23:23.501299] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:28.407 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:28.408 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:28.408 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:28.408 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:28.408 20:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:28.408 20:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:28.408 20:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:28.408 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:28.408 "name": "Existed_Raid", 00:24:28.408 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:28.408 "strip_size_kb": 64, 00:24:28.408 "state": "configuring", 00:24:28.408 "raid_level": "concat", 00:24:28.408 "superblock": false, 00:24:28.408 "num_base_bdevs": 4, 00:24:28.408 "num_base_bdevs_discovered": 3, 00:24:28.408 "num_base_bdevs_operational": 4, 00:24:28.408 "base_bdevs_list": [ 00:24:28.408 { 00:24:28.408 "name": "BaseBdev1", 00:24:28.408 "uuid": "9945eea1-903a-4100-bcdc-514dbb8ba27e", 00:24:28.408 "is_configured": true, 00:24:28.408 "data_offset": 0, 00:24:28.408 "data_size": 65536 00:24:28.408 }, 00:24:28.408 { 00:24:28.408 "name": null, 00:24:28.408 "uuid": "138795bb-0cdb-4865-8dac-f5c15361950f", 00:24:28.408 "is_configured": false, 00:24:28.408 "data_offset": 0, 00:24:28.408 "data_size": 65536 00:24:28.408 }, 00:24:28.408 { 00:24:28.408 "name": "BaseBdev3", 00:24:28.408 "uuid": "03b4c4db-4b3c-4fbf-a828-636d274d1e90", 00:24:28.408 "is_configured": true, 00:24:28.408 "data_offset": 0, 00:24:28.408 "data_size": 65536 00:24:28.408 }, 00:24:28.408 { 00:24:28.408 "name": "BaseBdev4", 00:24:28.408 "uuid": "4931312f-2000-4f27-b4a3-6a3890219c86", 00:24:28.408 "is_configured": true, 00:24:28.408 "data_offset": 0, 00:24:28.408 "data_size": 65536 00:24:28.408 } 00:24:28.408 ] 00:24:28.408 }' 00:24:28.408 20:23:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:28.408 20:23:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:28.973 [2024-10-01 20:23:24.057473] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:28.973 "name": "Existed_Raid", 00:24:28.973 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:28.973 "strip_size_kb": 64, 00:24:28.973 "state": "configuring", 00:24:28.973 "raid_level": "concat", 00:24:28.973 "superblock": false, 00:24:28.973 "num_base_bdevs": 4, 00:24:28.973 "num_base_bdevs_discovered": 2, 00:24:28.973 "num_base_bdevs_operational": 4, 00:24:28.973 "base_bdevs_list": [ 00:24:28.973 { 00:24:28.973 "name": null, 00:24:28.973 "uuid": "9945eea1-903a-4100-bcdc-514dbb8ba27e", 00:24:28.973 "is_configured": false, 00:24:28.973 "data_offset": 0, 00:24:28.973 "data_size": 65536 00:24:28.973 }, 00:24:28.973 { 00:24:28.973 "name": null, 00:24:28.973 "uuid": "138795bb-0cdb-4865-8dac-f5c15361950f", 00:24:28.973 "is_configured": false, 00:24:28.973 "data_offset": 0, 00:24:28.973 "data_size": 65536 00:24:28.973 }, 00:24:28.973 { 00:24:28.973 "name": "BaseBdev3", 00:24:28.973 "uuid": "03b4c4db-4b3c-4fbf-a828-636d274d1e90", 00:24:28.973 "is_configured": true, 00:24:28.973 "data_offset": 0, 00:24:28.973 "data_size": 65536 00:24:28.973 }, 00:24:28.973 { 00:24:28.973 "name": "BaseBdev4", 00:24:28.973 "uuid": "4931312f-2000-4f27-b4a3-6a3890219c86", 00:24:28.973 "is_configured": true, 00:24:28.973 "data_offset": 0, 00:24:28.973 "data_size": 65536 00:24:28.973 } 00:24:28.973 ] 00:24:28.973 }' 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:28.973 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:29.540 [2024-10-01 20:23:24.719891] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:29.540 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:29.540 "name": "Existed_Raid", 00:24:29.540 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:29.540 "strip_size_kb": 64, 00:24:29.540 "state": "configuring", 00:24:29.540 "raid_level": "concat", 00:24:29.540 "superblock": false, 00:24:29.540 "num_base_bdevs": 4, 00:24:29.540 "num_base_bdevs_discovered": 3, 00:24:29.540 "num_base_bdevs_operational": 4, 00:24:29.540 "base_bdevs_list": [ 00:24:29.540 { 00:24:29.540 "name": null, 00:24:29.540 "uuid": "9945eea1-903a-4100-bcdc-514dbb8ba27e", 00:24:29.540 "is_configured": false, 00:24:29.540 "data_offset": 0, 00:24:29.540 "data_size": 65536 00:24:29.540 }, 00:24:29.540 { 00:24:29.540 "name": "BaseBdev2", 00:24:29.540 "uuid": "138795bb-0cdb-4865-8dac-f5c15361950f", 00:24:29.540 "is_configured": true, 00:24:29.540 "data_offset": 0, 00:24:29.540 "data_size": 65536 00:24:29.540 }, 00:24:29.540 { 00:24:29.540 "name": "BaseBdev3", 00:24:29.540 "uuid": "03b4c4db-4b3c-4fbf-a828-636d274d1e90", 00:24:29.540 "is_configured": true, 00:24:29.540 "data_offset": 0, 00:24:29.540 "data_size": 65536 00:24:29.540 }, 00:24:29.540 { 00:24:29.540 "name": "BaseBdev4", 00:24:29.540 "uuid": "4931312f-2000-4f27-b4a3-6a3890219c86", 00:24:29.541 "is_configured": true, 00:24:29.541 "data_offset": 0, 00:24:29.541 "data_size": 65536 00:24:29.541 } 00:24:29.541 ] 00:24:29.541 }' 00:24:29.541 20:23:24 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:29.541 20:23:24 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:30.107 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:30.107 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:30.107 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:30.107 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:24:30.107 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:30.107 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:24:30.107 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:30.107 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:30.107 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:30.107 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:24:30.107 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:30.107 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 9945eea1-903a-4100-bcdc-514dbb8ba27e 00:24:30.107 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:30.107 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:30.365 [2024-10-01 20:23:25.380236] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:24:30.365 [2024-10-01 20:23:25.380311] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:24:30.365 [2024-10-01 20:23:25.380325] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:24:30.365 [2024-10-01 20:23:25.380664] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:24:30.365 [2024-10-01 20:23:25.380884] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:24:30.365 [2024-10-01 20:23:25.380906] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:24:30.365 NewBaseBdev 00:24:30.365 [2024-10-01 20:23:25.381218] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:30.365 [ 00:24:30.365 { 00:24:30.365 "name": "NewBaseBdev", 00:24:30.365 "aliases": [ 00:24:30.365 "9945eea1-903a-4100-bcdc-514dbb8ba27e" 00:24:30.365 ], 00:24:30.365 "product_name": "Malloc disk", 00:24:30.365 "block_size": 512, 00:24:30.365 "num_blocks": 65536, 00:24:30.365 "uuid": "9945eea1-903a-4100-bcdc-514dbb8ba27e", 00:24:30.365 "assigned_rate_limits": { 00:24:30.365 "rw_ios_per_sec": 0, 00:24:30.365 "rw_mbytes_per_sec": 0, 00:24:30.365 "r_mbytes_per_sec": 0, 00:24:30.365 "w_mbytes_per_sec": 0 00:24:30.365 }, 00:24:30.365 "claimed": true, 00:24:30.365 "claim_type": "exclusive_write", 00:24:30.365 "zoned": false, 00:24:30.365 "supported_io_types": { 00:24:30.365 "read": true, 00:24:30.365 "write": true, 00:24:30.365 "unmap": true, 00:24:30.365 "flush": true, 00:24:30.365 "reset": true, 00:24:30.365 "nvme_admin": false, 00:24:30.365 "nvme_io": false, 00:24:30.365 "nvme_io_md": false, 00:24:30.365 "write_zeroes": true, 00:24:30.365 "zcopy": true, 00:24:30.365 "get_zone_info": false, 00:24:30.365 "zone_management": false, 00:24:30.365 "zone_append": false, 00:24:30.365 "compare": false, 00:24:30.365 "compare_and_write": false, 00:24:30.365 "abort": true, 00:24:30.365 "seek_hole": false, 00:24:30.365 "seek_data": false, 00:24:30.365 "copy": true, 00:24:30.365 "nvme_iov_md": false 00:24:30.365 }, 00:24:30.365 "memory_domains": [ 00:24:30.365 { 00:24:30.365 "dma_device_id": "system", 00:24:30.365 "dma_device_type": 1 00:24:30.365 }, 00:24:30.365 { 00:24:30.365 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:30.365 "dma_device_type": 2 00:24:30.365 } 00:24:30.365 ], 00:24:30.365 "driver_specific": {} 00:24:30.365 } 00:24:30.365 ] 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:30.365 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:30.365 "name": "Existed_Raid", 00:24:30.365 "uuid": "59523538-0a20-4e28-be18-2006746e2157", 00:24:30.365 "strip_size_kb": 64, 00:24:30.365 "state": "online", 00:24:30.365 "raid_level": "concat", 00:24:30.365 "superblock": false, 00:24:30.365 "num_base_bdevs": 4, 00:24:30.365 "num_base_bdevs_discovered": 4, 00:24:30.365 "num_base_bdevs_operational": 4, 00:24:30.365 "base_bdevs_list": [ 00:24:30.365 { 00:24:30.365 "name": "NewBaseBdev", 00:24:30.365 "uuid": "9945eea1-903a-4100-bcdc-514dbb8ba27e", 00:24:30.365 "is_configured": true, 00:24:30.365 "data_offset": 0, 00:24:30.365 "data_size": 65536 00:24:30.365 }, 00:24:30.365 { 00:24:30.365 "name": "BaseBdev2", 00:24:30.365 "uuid": "138795bb-0cdb-4865-8dac-f5c15361950f", 00:24:30.365 "is_configured": true, 00:24:30.365 "data_offset": 0, 00:24:30.365 "data_size": 65536 00:24:30.365 }, 00:24:30.365 { 00:24:30.365 "name": "BaseBdev3", 00:24:30.365 "uuid": "03b4c4db-4b3c-4fbf-a828-636d274d1e90", 00:24:30.365 "is_configured": true, 00:24:30.365 "data_offset": 0, 00:24:30.365 "data_size": 65536 00:24:30.365 }, 00:24:30.365 { 00:24:30.365 "name": "BaseBdev4", 00:24:30.365 "uuid": "4931312f-2000-4f27-b4a3-6a3890219c86", 00:24:30.365 "is_configured": true, 00:24:30.366 "data_offset": 0, 00:24:30.366 "data_size": 65536 00:24:30.366 } 00:24:30.366 ] 00:24:30.366 }' 00:24:30.366 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:30.366 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:30.933 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:24:30.933 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:24:30.933 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:24:30.933 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:24:30.933 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:24:30.933 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:24:30.933 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:24:30.933 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:24:30.933 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:30.933 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:30.933 [2024-10-01 20:23:25.932967] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:24:30.933 20:23:25 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:30.933 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:30.933 "name": "Existed_Raid", 00:24:30.933 "aliases": [ 00:24:30.933 "59523538-0a20-4e28-be18-2006746e2157" 00:24:30.933 ], 00:24:30.933 "product_name": "Raid Volume", 00:24:30.933 "block_size": 512, 00:24:30.933 "num_blocks": 262144, 00:24:30.933 "uuid": "59523538-0a20-4e28-be18-2006746e2157", 00:24:30.933 "assigned_rate_limits": { 00:24:30.933 "rw_ios_per_sec": 0, 00:24:30.933 "rw_mbytes_per_sec": 0, 00:24:30.933 "r_mbytes_per_sec": 0, 00:24:30.933 "w_mbytes_per_sec": 0 00:24:30.933 }, 00:24:30.933 "claimed": false, 00:24:30.933 "zoned": false, 00:24:30.933 "supported_io_types": { 00:24:30.933 "read": true, 00:24:30.933 "write": true, 00:24:30.933 "unmap": true, 00:24:30.933 "flush": true, 00:24:30.933 "reset": true, 00:24:30.933 "nvme_admin": false, 00:24:30.933 "nvme_io": false, 00:24:30.933 "nvme_io_md": false, 00:24:30.933 "write_zeroes": true, 00:24:30.933 "zcopy": false, 00:24:30.933 "get_zone_info": false, 00:24:30.933 "zone_management": false, 00:24:30.933 "zone_append": false, 00:24:30.933 "compare": false, 00:24:30.934 "compare_and_write": false, 00:24:30.934 "abort": false, 00:24:30.934 "seek_hole": false, 00:24:30.934 "seek_data": false, 00:24:30.934 "copy": false, 00:24:30.934 "nvme_iov_md": false 00:24:30.934 }, 00:24:30.934 "memory_domains": [ 00:24:30.934 { 00:24:30.934 "dma_device_id": "system", 00:24:30.934 "dma_device_type": 1 00:24:30.934 }, 00:24:30.934 { 00:24:30.934 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:30.934 "dma_device_type": 2 00:24:30.934 }, 00:24:30.934 { 00:24:30.934 "dma_device_id": "system", 00:24:30.934 "dma_device_type": 1 00:24:30.934 }, 00:24:30.934 { 00:24:30.934 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:30.934 "dma_device_type": 2 00:24:30.934 }, 00:24:30.934 { 00:24:30.934 "dma_device_id": "system", 00:24:30.934 "dma_device_type": 1 00:24:30.934 }, 00:24:30.934 { 00:24:30.934 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:30.934 "dma_device_type": 2 00:24:30.934 }, 00:24:30.934 { 00:24:30.934 "dma_device_id": "system", 00:24:30.934 "dma_device_type": 1 00:24:30.934 }, 00:24:30.934 { 00:24:30.934 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:30.934 "dma_device_type": 2 00:24:30.934 } 00:24:30.934 ], 00:24:30.934 "driver_specific": { 00:24:30.934 "raid": { 00:24:30.934 "uuid": "59523538-0a20-4e28-be18-2006746e2157", 00:24:30.934 "strip_size_kb": 64, 00:24:30.934 "state": "online", 00:24:30.934 "raid_level": "concat", 00:24:30.934 "superblock": false, 00:24:30.934 "num_base_bdevs": 4, 00:24:30.934 "num_base_bdevs_discovered": 4, 00:24:30.934 "num_base_bdevs_operational": 4, 00:24:30.934 "base_bdevs_list": [ 00:24:30.934 { 00:24:30.934 "name": "NewBaseBdev", 00:24:30.934 "uuid": "9945eea1-903a-4100-bcdc-514dbb8ba27e", 00:24:30.934 "is_configured": true, 00:24:30.934 "data_offset": 0, 00:24:30.934 "data_size": 65536 00:24:30.934 }, 00:24:30.934 { 00:24:30.934 "name": "BaseBdev2", 00:24:30.934 "uuid": "138795bb-0cdb-4865-8dac-f5c15361950f", 00:24:30.934 "is_configured": true, 00:24:30.934 "data_offset": 0, 00:24:30.934 "data_size": 65536 00:24:30.934 }, 00:24:30.934 { 00:24:30.934 "name": "BaseBdev3", 00:24:30.934 "uuid": "03b4c4db-4b3c-4fbf-a828-636d274d1e90", 00:24:30.934 "is_configured": true, 00:24:30.934 "data_offset": 0, 00:24:30.934 "data_size": 65536 00:24:30.934 }, 00:24:30.934 { 00:24:30.934 "name": "BaseBdev4", 00:24:30.934 "uuid": "4931312f-2000-4f27-b4a3-6a3890219c86", 00:24:30.934 "is_configured": true, 00:24:30.934 "data_offset": 0, 00:24:30.934 "data_size": 65536 00:24:30.934 } 00:24:30.934 ] 00:24:30.934 } 00:24:30.934 } 00:24:30.934 }' 00:24:30.934 20:23:25 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:24:30.934 BaseBdev2 00:24:30.934 BaseBdev3 00:24:30.934 BaseBdev4' 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:30.934 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:31.193 [2024-10-01 20:23:26.320600] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:24:31.193 [2024-10-01 20:23:26.320816] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:24:31.193 [2024-10-01 20:23:26.321073] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:24:31.193 [2024-10-01 20:23:26.321306] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:24:31.193 [2024-10-01 20:23:26.321335] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 72059 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 72059 ']' 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 72059 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72059 00:24:31.193 killing process with pid 72059 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72059' 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 72059 00:24:31.193 [2024-10-01 20:23:26.359756] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:24:31.193 20:23:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 72059 00:24:31.758 [2024-10-01 20:23:26.734383] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:24:33.660 00:24:33.660 real 0m13.910s 00:24:33.660 user 0m22.392s 00:24:33.660 sys 0m1.926s 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:24:33.660 ************************************ 00:24:33.660 END TEST raid_state_function_test 00:24:33.660 ************************************ 00:24:33.660 20:23:28 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 4 true 00:24:33.660 20:23:28 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:24:33.660 20:23:28 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:24:33.660 20:23:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:24:33.660 ************************************ 00:24:33.660 START TEST raid_state_function_test_sb 00:24:33.660 ************************************ 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 4 true 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=72753 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:24:33.660 Process raid pid: 72753 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72753' 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 72753 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 72753 ']' 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:33.660 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:33.660 20:23:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:33.660 [2024-10-01 20:23:28.755253] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:24:33.660 [2024-10-01 20:23:28.755448] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:33.919 [2024-10-01 20:23:28.935199] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:34.179 [2024-10-01 20:23:29.186012] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:24:34.179 [2024-10-01 20:23:29.402842] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:34.179 [2024-10-01 20:23:29.402932] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:34.748 [2024-10-01 20:23:29.835557] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:24:34.748 [2024-10-01 20:23:29.835846] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:24:34.748 [2024-10-01 20:23:29.835978] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:24:34.748 [2024-10-01 20:23:29.836042] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:24:34.748 [2024-10-01 20:23:29.836144] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:24:34.748 [2024-10-01 20:23:29.836215] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:24:34.748 [2024-10-01 20:23:29.836326] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:24:34.748 [2024-10-01 20:23:29.836385] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:34.748 "name": "Existed_Raid", 00:24:34.748 "uuid": "0d12bc9e-eedd-4ecd-8ebe-5a3cdb0f055c", 00:24:34.748 "strip_size_kb": 64, 00:24:34.748 "state": "configuring", 00:24:34.748 "raid_level": "concat", 00:24:34.748 "superblock": true, 00:24:34.748 "num_base_bdevs": 4, 00:24:34.748 "num_base_bdevs_discovered": 0, 00:24:34.748 "num_base_bdevs_operational": 4, 00:24:34.748 "base_bdevs_list": [ 00:24:34.748 { 00:24:34.748 "name": "BaseBdev1", 00:24:34.748 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:34.748 "is_configured": false, 00:24:34.748 "data_offset": 0, 00:24:34.748 "data_size": 0 00:24:34.748 }, 00:24:34.748 { 00:24:34.748 "name": "BaseBdev2", 00:24:34.748 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:34.748 "is_configured": false, 00:24:34.748 "data_offset": 0, 00:24:34.748 "data_size": 0 00:24:34.748 }, 00:24:34.748 { 00:24:34.748 "name": "BaseBdev3", 00:24:34.748 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:34.748 "is_configured": false, 00:24:34.748 "data_offset": 0, 00:24:34.748 "data_size": 0 00:24:34.748 }, 00:24:34.748 { 00:24:34.748 "name": "BaseBdev4", 00:24:34.748 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:34.748 "is_configured": false, 00:24:34.748 "data_offset": 0, 00:24:34.748 "data_size": 0 00:24:34.748 } 00:24:34.748 ] 00:24:34.748 }' 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:34.748 20:23:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:35.316 [2024-10-01 20:23:30.363594] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:24:35.316 [2024-10-01 20:23:30.363804] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:35.316 [2024-10-01 20:23:30.371625] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:24:35.316 [2024-10-01 20:23:30.371818] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:24:35.316 [2024-10-01 20:23:30.371931] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:24:35.316 [2024-10-01 20:23:30.371992] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:24:35.316 [2024-10-01 20:23:30.372082] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:24:35.316 [2024-10-01 20:23:30.372139] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:24:35.316 [2024-10-01 20:23:30.372255] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:24:35.316 [2024-10-01 20:23:30.372312] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:35.316 [2024-10-01 20:23:30.417193] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:24:35.316 BaseBdev1 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:35.316 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:35.317 [ 00:24:35.317 { 00:24:35.317 "name": "BaseBdev1", 00:24:35.317 "aliases": [ 00:24:35.317 "a1a0ec63-627d-442e-9825-c697e6e1ae27" 00:24:35.317 ], 00:24:35.317 "product_name": "Malloc disk", 00:24:35.317 "block_size": 512, 00:24:35.317 "num_blocks": 65536, 00:24:35.317 "uuid": "a1a0ec63-627d-442e-9825-c697e6e1ae27", 00:24:35.317 "assigned_rate_limits": { 00:24:35.317 "rw_ios_per_sec": 0, 00:24:35.317 "rw_mbytes_per_sec": 0, 00:24:35.317 "r_mbytes_per_sec": 0, 00:24:35.317 "w_mbytes_per_sec": 0 00:24:35.317 }, 00:24:35.317 "claimed": true, 00:24:35.317 "claim_type": "exclusive_write", 00:24:35.317 "zoned": false, 00:24:35.317 "supported_io_types": { 00:24:35.317 "read": true, 00:24:35.317 "write": true, 00:24:35.317 "unmap": true, 00:24:35.317 "flush": true, 00:24:35.317 "reset": true, 00:24:35.317 "nvme_admin": false, 00:24:35.317 "nvme_io": false, 00:24:35.317 "nvme_io_md": false, 00:24:35.317 "write_zeroes": true, 00:24:35.317 "zcopy": true, 00:24:35.317 "get_zone_info": false, 00:24:35.317 "zone_management": false, 00:24:35.317 "zone_append": false, 00:24:35.317 "compare": false, 00:24:35.317 "compare_and_write": false, 00:24:35.317 "abort": true, 00:24:35.317 "seek_hole": false, 00:24:35.317 "seek_data": false, 00:24:35.317 "copy": true, 00:24:35.317 "nvme_iov_md": false 00:24:35.317 }, 00:24:35.317 "memory_domains": [ 00:24:35.317 { 00:24:35.317 "dma_device_id": "system", 00:24:35.317 "dma_device_type": 1 00:24:35.317 }, 00:24:35.317 { 00:24:35.317 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:35.317 "dma_device_type": 2 00:24:35.317 } 00:24:35.317 ], 00:24:35.317 "driver_specific": {} 00:24:35.317 } 00:24:35.317 ] 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:35.317 "name": "Existed_Raid", 00:24:35.317 "uuid": "7aaf8b19-d2ca-4922-8cab-00ea96447f31", 00:24:35.317 "strip_size_kb": 64, 00:24:35.317 "state": "configuring", 00:24:35.317 "raid_level": "concat", 00:24:35.317 "superblock": true, 00:24:35.317 "num_base_bdevs": 4, 00:24:35.317 "num_base_bdevs_discovered": 1, 00:24:35.317 "num_base_bdevs_operational": 4, 00:24:35.317 "base_bdevs_list": [ 00:24:35.317 { 00:24:35.317 "name": "BaseBdev1", 00:24:35.317 "uuid": "a1a0ec63-627d-442e-9825-c697e6e1ae27", 00:24:35.317 "is_configured": true, 00:24:35.317 "data_offset": 2048, 00:24:35.317 "data_size": 63488 00:24:35.317 }, 00:24:35.317 { 00:24:35.317 "name": "BaseBdev2", 00:24:35.317 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:35.317 "is_configured": false, 00:24:35.317 "data_offset": 0, 00:24:35.317 "data_size": 0 00:24:35.317 }, 00:24:35.317 { 00:24:35.317 "name": "BaseBdev3", 00:24:35.317 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:35.317 "is_configured": false, 00:24:35.317 "data_offset": 0, 00:24:35.317 "data_size": 0 00:24:35.317 }, 00:24:35.317 { 00:24:35.317 "name": "BaseBdev4", 00:24:35.317 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:35.317 "is_configured": false, 00:24:35.317 "data_offset": 0, 00:24:35.317 "data_size": 0 00:24:35.317 } 00:24:35.317 ] 00:24:35.317 }' 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:35.317 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:35.885 [2024-10-01 20:23:30.965484] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:24:35.885 [2024-10-01 20:23:30.965606] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:35.885 [2024-10-01 20:23:30.973496] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:24:35.885 [2024-10-01 20:23:30.976406] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:24:35.885 [2024-10-01 20:23:30.976632] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:24:35.885 [2024-10-01 20:23:30.976851] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:24:35.885 [2024-10-01 20:23:30.976889] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:24:35.885 [2024-10-01 20:23:30.976903] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:24:35.885 [2024-10-01 20:23:30.976918] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:35.885 20:23:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:35.885 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:35.885 "name": "Existed_Raid", 00:24:35.885 "uuid": "11485b50-3edc-4d03-93e7-f7974f109b4e", 00:24:35.885 "strip_size_kb": 64, 00:24:35.885 "state": "configuring", 00:24:35.885 "raid_level": "concat", 00:24:35.885 "superblock": true, 00:24:35.885 "num_base_bdevs": 4, 00:24:35.885 "num_base_bdevs_discovered": 1, 00:24:35.885 "num_base_bdevs_operational": 4, 00:24:35.885 "base_bdevs_list": [ 00:24:35.885 { 00:24:35.885 "name": "BaseBdev1", 00:24:35.885 "uuid": "a1a0ec63-627d-442e-9825-c697e6e1ae27", 00:24:35.885 "is_configured": true, 00:24:35.885 "data_offset": 2048, 00:24:35.885 "data_size": 63488 00:24:35.885 }, 00:24:35.885 { 00:24:35.885 "name": "BaseBdev2", 00:24:35.885 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:35.885 "is_configured": false, 00:24:35.885 "data_offset": 0, 00:24:35.885 "data_size": 0 00:24:35.885 }, 00:24:35.885 { 00:24:35.885 "name": "BaseBdev3", 00:24:35.885 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:35.885 "is_configured": false, 00:24:35.885 "data_offset": 0, 00:24:35.885 "data_size": 0 00:24:35.885 }, 00:24:35.885 { 00:24:35.886 "name": "BaseBdev4", 00:24:35.886 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:35.886 "is_configured": false, 00:24:35.886 "data_offset": 0, 00:24:35.886 "data_size": 0 00:24:35.886 } 00:24:35.886 ] 00:24:35.886 }' 00:24:35.886 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:35.886 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:36.454 [2024-10-01 20:23:31.538376] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:24:36.454 BaseBdev2 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:36.454 [ 00:24:36.454 { 00:24:36.454 "name": "BaseBdev2", 00:24:36.454 "aliases": [ 00:24:36.454 "b06947d2-e3bc-4198-b11b-603674dfcabe" 00:24:36.454 ], 00:24:36.454 "product_name": "Malloc disk", 00:24:36.454 "block_size": 512, 00:24:36.454 "num_blocks": 65536, 00:24:36.454 "uuid": "b06947d2-e3bc-4198-b11b-603674dfcabe", 00:24:36.454 "assigned_rate_limits": { 00:24:36.454 "rw_ios_per_sec": 0, 00:24:36.454 "rw_mbytes_per_sec": 0, 00:24:36.454 "r_mbytes_per_sec": 0, 00:24:36.454 "w_mbytes_per_sec": 0 00:24:36.454 }, 00:24:36.454 "claimed": true, 00:24:36.454 "claim_type": "exclusive_write", 00:24:36.454 "zoned": false, 00:24:36.454 "supported_io_types": { 00:24:36.454 "read": true, 00:24:36.454 "write": true, 00:24:36.454 "unmap": true, 00:24:36.454 "flush": true, 00:24:36.454 "reset": true, 00:24:36.454 "nvme_admin": false, 00:24:36.454 "nvme_io": false, 00:24:36.454 "nvme_io_md": false, 00:24:36.454 "write_zeroes": true, 00:24:36.454 "zcopy": true, 00:24:36.454 "get_zone_info": false, 00:24:36.454 "zone_management": false, 00:24:36.454 "zone_append": false, 00:24:36.454 "compare": false, 00:24:36.454 "compare_and_write": false, 00:24:36.454 "abort": true, 00:24:36.454 "seek_hole": false, 00:24:36.454 "seek_data": false, 00:24:36.454 "copy": true, 00:24:36.454 "nvme_iov_md": false 00:24:36.454 }, 00:24:36.454 "memory_domains": [ 00:24:36.454 { 00:24:36.454 "dma_device_id": "system", 00:24:36.454 "dma_device_type": 1 00:24:36.454 }, 00:24:36.454 { 00:24:36.454 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:36.454 "dma_device_type": 2 00:24:36.454 } 00:24:36.454 ], 00:24:36.454 "driver_specific": {} 00:24:36.454 } 00:24:36.454 ] 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:36.454 "name": "Existed_Raid", 00:24:36.454 "uuid": "11485b50-3edc-4d03-93e7-f7974f109b4e", 00:24:36.454 "strip_size_kb": 64, 00:24:36.454 "state": "configuring", 00:24:36.454 "raid_level": "concat", 00:24:36.454 "superblock": true, 00:24:36.454 "num_base_bdevs": 4, 00:24:36.454 "num_base_bdevs_discovered": 2, 00:24:36.454 "num_base_bdevs_operational": 4, 00:24:36.454 "base_bdevs_list": [ 00:24:36.454 { 00:24:36.454 "name": "BaseBdev1", 00:24:36.454 "uuid": "a1a0ec63-627d-442e-9825-c697e6e1ae27", 00:24:36.454 "is_configured": true, 00:24:36.454 "data_offset": 2048, 00:24:36.454 "data_size": 63488 00:24:36.454 }, 00:24:36.454 { 00:24:36.454 "name": "BaseBdev2", 00:24:36.454 "uuid": "b06947d2-e3bc-4198-b11b-603674dfcabe", 00:24:36.454 "is_configured": true, 00:24:36.454 "data_offset": 2048, 00:24:36.454 "data_size": 63488 00:24:36.454 }, 00:24:36.454 { 00:24:36.454 "name": "BaseBdev3", 00:24:36.454 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:36.454 "is_configured": false, 00:24:36.454 "data_offset": 0, 00:24:36.454 "data_size": 0 00:24:36.454 }, 00:24:36.454 { 00:24:36.454 "name": "BaseBdev4", 00:24:36.454 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:36.454 "is_configured": false, 00:24:36.454 "data_offset": 0, 00:24:36.454 "data_size": 0 00:24:36.454 } 00:24:36.454 ] 00:24:36.454 }' 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:36.454 20:23:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:37.023 [2024-10-01 20:23:32.126649] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:24:37.023 BaseBdev3 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:37.023 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:37.023 [ 00:24:37.023 { 00:24:37.023 "name": "BaseBdev3", 00:24:37.024 "aliases": [ 00:24:37.024 "cb197dd1-9cb2-4554-8aa5-81bb6534c78e" 00:24:37.024 ], 00:24:37.024 "product_name": "Malloc disk", 00:24:37.024 "block_size": 512, 00:24:37.024 "num_blocks": 65536, 00:24:37.024 "uuid": "cb197dd1-9cb2-4554-8aa5-81bb6534c78e", 00:24:37.024 "assigned_rate_limits": { 00:24:37.024 "rw_ios_per_sec": 0, 00:24:37.024 "rw_mbytes_per_sec": 0, 00:24:37.024 "r_mbytes_per_sec": 0, 00:24:37.024 "w_mbytes_per_sec": 0 00:24:37.024 }, 00:24:37.024 "claimed": true, 00:24:37.024 "claim_type": "exclusive_write", 00:24:37.024 "zoned": false, 00:24:37.024 "supported_io_types": { 00:24:37.024 "read": true, 00:24:37.024 "write": true, 00:24:37.024 "unmap": true, 00:24:37.024 "flush": true, 00:24:37.024 "reset": true, 00:24:37.024 "nvme_admin": false, 00:24:37.024 "nvme_io": false, 00:24:37.024 "nvme_io_md": false, 00:24:37.024 "write_zeroes": true, 00:24:37.024 "zcopy": true, 00:24:37.024 "get_zone_info": false, 00:24:37.024 "zone_management": false, 00:24:37.024 "zone_append": false, 00:24:37.024 "compare": false, 00:24:37.024 "compare_and_write": false, 00:24:37.024 "abort": true, 00:24:37.024 "seek_hole": false, 00:24:37.024 "seek_data": false, 00:24:37.024 "copy": true, 00:24:37.024 "nvme_iov_md": false 00:24:37.024 }, 00:24:37.024 "memory_domains": [ 00:24:37.024 { 00:24:37.024 "dma_device_id": "system", 00:24:37.024 "dma_device_type": 1 00:24:37.024 }, 00:24:37.024 { 00:24:37.024 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:37.024 "dma_device_type": 2 00:24:37.024 } 00:24:37.024 ], 00:24:37.024 "driver_specific": {} 00:24:37.024 } 00:24:37.024 ] 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:37.024 "name": "Existed_Raid", 00:24:37.024 "uuid": "11485b50-3edc-4d03-93e7-f7974f109b4e", 00:24:37.024 "strip_size_kb": 64, 00:24:37.024 "state": "configuring", 00:24:37.024 "raid_level": "concat", 00:24:37.024 "superblock": true, 00:24:37.024 "num_base_bdevs": 4, 00:24:37.024 "num_base_bdevs_discovered": 3, 00:24:37.024 "num_base_bdevs_operational": 4, 00:24:37.024 "base_bdevs_list": [ 00:24:37.024 { 00:24:37.024 "name": "BaseBdev1", 00:24:37.024 "uuid": "a1a0ec63-627d-442e-9825-c697e6e1ae27", 00:24:37.024 "is_configured": true, 00:24:37.024 "data_offset": 2048, 00:24:37.024 "data_size": 63488 00:24:37.024 }, 00:24:37.024 { 00:24:37.024 "name": "BaseBdev2", 00:24:37.024 "uuid": "b06947d2-e3bc-4198-b11b-603674dfcabe", 00:24:37.024 "is_configured": true, 00:24:37.024 "data_offset": 2048, 00:24:37.024 "data_size": 63488 00:24:37.024 }, 00:24:37.024 { 00:24:37.024 "name": "BaseBdev3", 00:24:37.024 "uuid": "cb197dd1-9cb2-4554-8aa5-81bb6534c78e", 00:24:37.024 "is_configured": true, 00:24:37.024 "data_offset": 2048, 00:24:37.024 "data_size": 63488 00:24:37.024 }, 00:24:37.024 { 00:24:37.024 "name": "BaseBdev4", 00:24:37.024 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:37.024 "is_configured": false, 00:24:37.024 "data_offset": 0, 00:24:37.024 "data_size": 0 00:24:37.024 } 00:24:37.024 ] 00:24:37.024 }' 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:37.024 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:37.620 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:24:37.620 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:37.620 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:37.620 [2024-10-01 20:23:32.725948] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:24:37.620 [2024-10-01 20:23:32.726590] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:24:37.620 [2024-10-01 20:23:32.726621] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:24:37.620 BaseBdev4 00:24:37.620 [2024-10-01 20:23:32.727049] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:24:37.620 [2024-10-01 20:23:32.727269] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:24:37.620 [2024-10-01 20:23:32.727300] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:24:37.620 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:37.620 [2024-10-01 20:23:32.727482] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:37.620 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:24:37.620 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:24:37.620 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:37.620 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:24:37.620 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:37.620 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:37.621 [ 00:24:37.621 { 00:24:37.621 "name": "BaseBdev4", 00:24:37.621 "aliases": [ 00:24:37.621 "fda80fce-3bef-466c-949f-93f546af4b7d" 00:24:37.621 ], 00:24:37.621 "product_name": "Malloc disk", 00:24:37.621 "block_size": 512, 00:24:37.621 "num_blocks": 65536, 00:24:37.621 "uuid": "fda80fce-3bef-466c-949f-93f546af4b7d", 00:24:37.621 "assigned_rate_limits": { 00:24:37.621 "rw_ios_per_sec": 0, 00:24:37.621 "rw_mbytes_per_sec": 0, 00:24:37.621 "r_mbytes_per_sec": 0, 00:24:37.621 "w_mbytes_per_sec": 0 00:24:37.621 }, 00:24:37.621 "claimed": true, 00:24:37.621 "claim_type": "exclusive_write", 00:24:37.621 "zoned": false, 00:24:37.621 "supported_io_types": { 00:24:37.621 "read": true, 00:24:37.621 "write": true, 00:24:37.621 "unmap": true, 00:24:37.621 "flush": true, 00:24:37.621 "reset": true, 00:24:37.621 "nvme_admin": false, 00:24:37.621 "nvme_io": false, 00:24:37.621 "nvme_io_md": false, 00:24:37.621 "write_zeroes": true, 00:24:37.621 "zcopy": true, 00:24:37.621 "get_zone_info": false, 00:24:37.621 "zone_management": false, 00:24:37.621 "zone_append": false, 00:24:37.621 "compare": false, 00:24:37.621 "compare_and_write": false, 00:24:37.621 "abort": true, 00:24:37.621 "seek_hole": false, 00:24:37.621 "seek_data": false, 00:24:37.621 "copy": true, 00:24:37.621 "nvme_iov_md": false 00:24:37.621 }, 00:24:37.621 "memory_domains": [ 00:24:37.621 { 00:24:37.621 "dma_device_id": "system", 00:24:37.621 "dma_device_type": 1 00:24:37.621 }, 00:24:37.621 { 00:24:37.621 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:37.621 "dma_device_type": 2 00:24:37.621 } 00:24:37.621 ], 00:24:37.621 "driver_specific": {} 00:24:37.621 } 00:24:37.621 ] 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:37.621 "name": "Existed_Raid", 00:24:37.621 "uuid": "11485b50-3edc-4d03-93e7-f7974f109b4e", 00:24:37.621 "strip_size_kb": 64, 00:24:37.621 "state": "online", 00:24:37.621 "raid_level": "concat", 00:24:37.621 "superblock": true, 00:24:37.621 "num_base_bdevs": 4, 00:24:37.621 "num_base_bdevs_discovered": 4, 00:24:37.621 "num_base_bdevs_operational": 4, 00:24:37.621 "base_bdevs_list": [ 00:24:37.621 { 00:24:37.621 "name": "BaseBdev1", 00:24:37.621 "uuid": "a1a0ec63-627d-442e-9825-c697e6e1ae27", 00:24:37.621 "is_configured": true, 00:24:37.621 "data_offset": 2048, 00:24:37.621 "data_size": 63488 00:24:37.621 }, 00:24:37.621 { 00:24:37.621 "name": "BaseBdev2", 00:24:37.621 "uuid": "b06947d2-e3bc-4198-b11b-603674dfcabe", 00:24:37.621 "is_configured": true, 00:24:37.621 "data_offset": 2048, 00:24:37.621 "data_size": 63488 00:24:37.621 }, 00:24:37.621 { 00:24:37.621 "name": "BaseBdev3", 00:24:37.621 "uuid": "cb197dd1-9cb2-4554-8aa5-81bb6534c78e", 00:24:37.621 "is_configured": true, 00:24:37.621 "data_offset": 2048, 00:24:37.621 "data_size": 63488 00:24:37.621 }, 00:24:37.621 { 00:24:37.621 "name": "BaseBdev4", 00:24:37.621 "uuid": "fda80fce-3bef-466c-949f-93f546af4b7d", 00:24:37.621 "is_configured": true, 00:24:37.621 "data_offset": 2048, 00:24:37.621 "data_size": 63488 00:24:37.621 } 00:24:37.621 ] 00:24:37.621 }' 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:37.621 20:23:32 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:38.199 [2024-10-01 20:23:33.290605] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:38.199 "name": "Existed_Raid", 00:24:38.199 "aliases": [ 00:24:38.199 "11485b50-3edc-4d03-93e7-f7974f109b4e" 00:24:38.199 ], 00:24:38.199 "product_name": "Raid Volume", 00:24:38.199 "block_size": 512, 00:24:38.199 "num_blocks": 253952, 00:24:38.199 "uuid": "11485b50-3edc-4d03-93e7-f7974f109b4e", 00:24:38.199 "assigned_rate_limits": { 00:24:38.199 "rw_ios_per_sec": 0, 00:24:38.199 "rw_mbytes_per_sec": 0, 00:24:38.199 "r_mbytes_per_sec": 0, 00:24:38.199 "w_mbytes_per_sec": 0 00:24:38.199 }, 00:24:38.199 "claimed": false, 00:24:38.199 "zoned": false, 00:24:38.199 "supported_io_types": { 00:24:38.199 "read": true, 00:24:38.199 "write": true, 00:24:38.199 "unmap": true, 00:24:38.199 "flush": true, 00:24:38.199 "reset": true, 00:24:38.199 "nvme_admin": false, 00:24:38.199 "nvme_io": false, 00:24:38.199 "nvme_io_md": false, 00:24:38.199 "write_zeroes": true, 00:24:38.199 "zcopy": false, 00:24:38.199 "get_zone_info": false, 00:24:38.199 "zone_management": false, 00:24:38.199 "zone_append": false, 00:24:38.199 "compare": false, 00:24:38.199 "compare_and_write": false, 00:24:38.199 "abort": false, 00:24:38.199 "seek_hole": false, 00:24:38.199 "seek_data": false, 00:24:38.199 "copy": false, 00:24:38.199 "nvme_iov_md": false 00:24:38.199 }, 00:24:38.199 "memory_domains": [ 00:24:38.199 { 00:24:38.199 "dma_device_id": "system", 00:24:38.199 "dma_device_type": 1 00:24:38.199 }, 00:24:38.199 { 00:24:38.199 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:38.199 "dma_device_type": 2 00:24:38.199 }, 00:24:38.199 { 00:24:38.199 "dma_device_id": "system", 00:24:38.199 "dma_device_type": 1 00:24:38.199 }, 00:24:38.199 { 00:24:38.199 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:38.199 "dma_device_type": 2 00:24:38.199 }, 00:24:38.199 { 00:24:38.199 "dma_device_id": "system", 00:24:38.199 "dma_device_type": 1 00:24:38.199 }, 00:24:38.199 { 00:24:38.199 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:38.199 "dma_device_type": 2 00:24:38.199 }, 00:24:38.199 { 00:24:38.199 "dma_device_id": "system", 00:24:38.199 "dma_device_type": 1 00:24:38.199 }, 00:24:38.199 { 00:24:38.199 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:38.199 "dma_device_type": 2 00:24:38.199 } 00:24:38.199 ], 00:24:38.199 "driver_specific": { 00:24:38.199 "raid": { 00:24:38.199 "uuid": "11485b50-3edc-4d03-93e7-f7974f109b4e", 00:24:38.199 "strip_size_kb": 64, 00:24:38.199 "state": "online", 00:24:38.199 "raid_level": "concat", 00:24:38.199 "superblock": true, 00:24:38.199 "num_base_bdevs": 4, 00:24:38.199 "num_base_bdevs_discovered": 4, 00:24:38.199 "num_base_bdevs_operational": 4, 00:24:38.199 "base_bdevs_list": [ 00:24:38.199 { 00:24:38.199 "name": "BaseBdev1", 00:24:38.199 "uuid": "a1a0ec63-627d-442e-9825-c697e6e1ae27", 00:24:38.199 "is_configured": true, 00:24:38.199 "data_offset": 2048, 00:24:38.199 "data_size": 63488 00:24:38.199 }, 00:24:38.199 { 00:24:38.199 "name": "BaseBdev2", 00:24:38.199 "uuid": "b06947d2-e3bc-4198-b11b-603674dfcabe", 00:24:38.199 "is_configured": true, 00:24:38.199 "data_offset": 2048, 00:24:38.199 "data_size": 63488 00:24:38.199 }, 00:24:38.199 { 00:24:38.199 "name": "BaseBdev3", 00:24:38.199 "uuid": "cb197dd1-9cb2-4554-8aa5-81bb6534c78e", 00:24:38.199 "is_configured": true, 00:24:38.199 "data_offset": 2048, 00:24:38.199 "data_size": 63488 00:24:38.199 }, 00:24:38.199 { 00:24:38.199 "name": "BaseBdev4", 00:24:38.199 "uuid": "fda80fce-3bef-466c-949f-93f546af4b7d", 00:24:38.199 "is_configured": true, 00:24:38.199 "data_offset": 2048, 00:24:38.199 "data_size": 63488 00:24:38.199 } 00:24:38.199 ] 00:24:38.199 } 00:24:38.199 } 00:24:38.199 }' 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:24:38.199 BaseBdev2 00:24:38.199 BaseBdev3 00:24:38.199 BaseBdev4' 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:38.199 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:38.459 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:38.459 [2024-10-01 20:23:33.654453] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:24:38.459 [2024-10-01 20:23:33.654666] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:24:38.459 [2024-10-01 20:23:33.654905] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:38.719 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:38.719 "name": "Existed_Raid", 00:24:38.719 "uuid": "11485b50-3edc-4d03-93e7-f7974f109b4e", 00:24:38.719 "strip_size_kb": 64, 00:24:38.719 "state": "offline", 00:24:38.719 "raid_level": "concat", 00:24:38.719 "superblock": true, 00:24:38.719 "num_base_bdevs": 4, 00:24:38.719 "num_base_bdevs_discovered": 3, 00:24:38.719 "num_base_bdevs_operational": 3, 00:24:38.719 "base_bdevs_list": [ 00:24:38.719 { 00:24:38.719 "name": null, 00:24:38.719 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:38.719 "is_configured": false, 00:24:38.719 "data_offset": 0, 00:24:38.720 "data_size": 63488 00:24:38.720 }, 00:24:38.720 { 00:24:38.720 "name": "BaseBdev2", 00:24:38.720 "uuid": "b06947d2-e3bc-4198-b11b-603674dfcabe", 00:24:38.720 "is_configured": true, 00:24:38.720 "data_offset": 2048, 00:24:38.720 "data_size": 63488 00:24:38.720 }, 00:24:38.720 { 00:24:38.720 "name": "BaseBdev3", 00:24:38.720 "uuid": "cb197dd1-9cb2-4554-8aa5-81bb6534c78e", 00:24:38.720 "is_configured": true, 00:24:38.720 "data_offset": 2048, 00:24:38.720 "data_size": 63488 00:24:38.720 }, 00:24:38.720 { 00:24:38.720 "name": "BaseBdev4", 00:24:38.720 "uuid": "fda80fce-3bef-466c-949f-93f546af4b7d", 00:24:38.720 "is_configured": true, 00:24:38.720 "data_offset": 2048, 00:24:38.720 "data_size": 63488 00:24:38.720 } 00:24:38.720 ] 00:24:38.720 }' 00:24:38.720 20:23:33 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:38.720 20:23:33 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.286 [2024-10-01 20:23:34.326606] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.286 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.286 [2024-10-01 20:23:34.463313] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.545 [2024-10-01 20:23:34.609411] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:24:39.545 [2024-10-01 20:23:34.609594] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.545 BaseBdev2 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.545 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:24:39.804 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:24:39.804 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:39.804 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:24:39.804 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:39.804 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:39.804 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:39.804 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.804 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.805 [ 00:24:39.805 { 00:24:39.805 "name": "BaseBdev2", 00:24:39.805 "aliases": [ 00:24:39.805 "7a85b44e-5096-4d78-980b-4fa5e92fcfc4" 00:24:39.805 ], 00:24:39.805 "product_name": "Malloc disk", 00:24:39.805 "block_size": 512, 00:24:39.805 "num_blocks": 65536, 00:24:39.805 "uuid": "7a85b44e-5096-4d78-980b-4fa5e92fcfc4", 00:24:39.805 "assigned_rate_limits": { 00:24:39.805 "rw_ios_per_sec": 0, 00:24:39.805 "rw_mbytes_per_sec": 0, 00:24:39.805 "r_mbytes_per_sec": 0, 00:24:39.805 "w_mbytes_per_sec": 0 00:24:39.805 }, 00:24:39.805 "claimed": false, 00:24:39.805 "zoned": false, 00:24:39.805 "supported_io_types": { 00:24:39.805 "read": true, 00:24:39.805 "write": true, 00:24:39.805 "unmap": true, 00:24:39.805 "flush": true, 00:24:39.805 "reset": true, 00:24:39.805 "nvme_admin": false, 00:24:39.805 "nvme_io": false, 00:24:39.805 "nvme_io_md": false, 00:24:39.805 "write_zeroes": true, 00:24:39.805 "zcopy": true, 00:24:39.805 "get_zone_info": false, 00:24:39.805 "zone_management": false, 00:24:39.805 "zone_append": false, 00:24:39.805 "compare": false, 00:24:39.805 "compare_and_write": false, 00:24:39.805 "abort": true, 00:24:39.805 "seek_hole": false, 00:24:39.805 "seek_data": false, 00:24:39.805 "copy": true, 00:24:39.805 "nvme_iov_md": false 00:24:39.805 }, 00:24:39.805 "memory_domains": [ 00:24:39.805 { 00:24:39.805 "dma_device_id": "system", 00:24:39.805 "dma_device_type": 1 00:24:39.805 }, 00:24:39.805 { 00:24:39.805 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:39.805 "dma_device_type": 2 00:24:39.805 } 00:24:39.805 ], 00:24:39.805 "driver_specific": {} 00:24:39.805 } 00:24:39.805 ] 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.805 BaseBdev3 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.805 [ 00:24:39.805 { 00:24:39.805 "name": "BaseBdev3", 00:24:39.805 "aliases": [ 00:24:39.805 "e1229a7c-c85d-4109-bcf3-7c232657dd40" 00:24:39.805 ], 00:24:39.805 "product_name": "Malloc disk", 00:24:39.805 "block_size": 512, 00:24:39.805 "num_blocks": 65536, 00:24:39.805 "uuid": "e1229a7c-c85d-4109-bcf3-7c232657dd40", 00:24:39.805 "assigned_rate_limits": { 00:24:39.805 "rw_ios_per_sec": 0, 00:24:39.805 "rw_mbytes_per_sec": 0, 00:24:39.805 "r_mbytes_per_sec": 0, 00:24:39.805 "w_mbytes_per_sec": 0 00:24:39.805 }, 00:24:39.805 "claimed": false, 00:24:39.805 "zoned": false, 00:24:39.805 "supported_io_types": { 00:24:39.805 "read": true, 00:24:39.805 "write": true, 00:24:39.805 "unmap": true, 00:24:39.805 "flush": true, 00:24:39.805 "reset": true, 00:24:39.805 "nvme_admin": false, 00:24:39.805 "nvme_io": false, 00:24:39.805 "nvme_io_md": false, 00:24:39.805 "write_zeroes": true, 00:24:39.805 "zcopy": true, 00:24:39.805 "get_zone_info": false, 00:24:39.805 "zone_management": false, 00:24:39.805 "zone_append": false, 00:24:39.805 "compare": false, 00:24:39.805 "compare_and_write": false, 00:24:39.805 "abort": true, 00:24:39.805 "seek_hole": false, 00:24:39.805 "seek_data": false, 00:24:39.805 "copy": true, 00:24:39.805 "nvme_iov_md": false 00:24:39.805 }, 00:24:39.805 "memory_domains": [ 00:24:39.805 { 00:24:39.805 "dma_device_id": "system", 00:24:39.805 "dma_device_type": 1 00:24:39.805 }, 00:24:39.805 { 00:24:39.805 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:39.805 "dma_device_type": 2 00:24:39.805 } 00:24:39.805 ], 00:24:39.805 "driver_specific": {} 00:24:39.805 } 00:24:39.805 ] 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.805 BaseBdev4 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.805 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.805 [ 00:24:39.805 { 00:24:39.805 "name": "BaseBdev4", 00:24:39.805 "aliases": [ 00:24:39.805 "289dece3-db1e-486d-a336-fdae863f6ff2" 00:24:39.805 ], 00:24:39.805 "product_name": "Malloc disk", 00:24:39.805 "block_size": 512, 00:24:39.805 "num_blocks": 65536, 00:24:39.805 "uuid": "289dece3-db1e-486d-a336-fdae863f6ff2", 00:24:39.805 "assigned_rate_limits": { 00:24:39.805 "rw_ios_per_sec": 0, 00:24:39.805 "rw_mbytes_per_sec": 0, 00:24:39.805 "r_mbytes_per_sec": 0, 00:24:39.805 "w_mbytes_per_sec": 0 00:24:39.805 }, 00:24:39.805 "claimed": false, 00:24:39.805 "zoned": false, 00:24:39.805 "supported_io_types": { 00:24:39.805 "read": true, 00:24:39.805 "write": true, 00:24:39.805 "unmap": true, 00:24:39.805 "flush": true, 00:24:39.805 "reset": true, 00:24:39.805 "nvme_admin": false, 00:24:39.806 "nvme_io": false, 00:24:39.806 "nvme_io_md": false, 00:24:39.806 "write_zeroes": true, 00:24:39.806 "zcopy": true, 00:24:39.806 "get_zone_info": false, 00:24:39.806 "zone_management": false, 00:24:39.806 "zone_append": false, 00:24:39.806 "compare": false, 00:24:39.806 "compare_and_write": false, 00:24:39.806 "abort": true, 00:24:39.806 "seek_hole": false, 00:24:39.806 "seek_data": false, 00:24:39.806 "copy": true, 00:24:39.806 "nvme_iov_md": false 00:24:39.806 }, 00:24:39.806 "memory_domains": [ 00:24:39.806 { 00:24:39.806 "dma_device_id": "system", 00:24:39.806 "dma_device_type": 1 00:24:39.806 }, 00:24:39.806 { 00:24:39.806 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:39.806 "dma_device_type": 2 00:24:39.806 } 00:24:39.806 ], 00:24:39.806 "driver_specific": {} 00:24:39.806 } 00:24:39.806 ] 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.806 [2024-10-01 20:23:34.975315] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:24:39.806 [2024-10-01 20:23:34.975374] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:24:39.806 [2024-10-01 20:23:34.975407] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:24:39.806 [2024-10-01 20:23:34.977984] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:24:39.806 [2024-10-01 20:23:34.978057] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:39.806 20:23:34 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:39.806 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:39.806 "name": "Existed_Raid", 00:24:39.806 "uuid": "b59a4974-e22f-43fd-8503-cde7d66e7ffc", 00:24:39.806 "strip_size_kb": 64, 00:24:39.806 "state": "configuring", 00:24:39.806 "raid_level": "concat", 00:24:39.806 "superblock": true, 00:24:39.806 "num_base_bdevs": 4, 00:24:39.806 "num_base_bdevs_discovered": 3, 00:24:39.806 "num_base_bdevs_operational": 4, 00:24:39.806 "base_bdevs_list": [ 00:24:39.806 { 00:24:39.806 "name": "BaseBdev1", 00:24:39.806 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:39.806 "is_configured": false, 00:24:39.806 "data_offset": 0, 00:24:39.806 "data_size": 0 00:24:39.806 }, 00:24:39.806 { 00:24:39.806 "name": "BaseBdev2", 00:24:39.806 "uuid": "7a85b44e-5096-4d78-980b-4fa5e92fcfc4", 00:24:39.806 "is_configured": true, 00:24:39.806 "data_offset": 2048, 00:24:39.806 "data_size": 63488 00:24:39.806 }, 00:24:39.806 { 00:24:39.806 "name": "BaseBdev3", 00:24:39.806 "uuid": "e1229a7c-c85d-4109-bcf3-7c232657dd40", 00:24:39.806 "is_configured": true, 00:24:39.806 "data_offset": 2048, 00:24:39.806 "data_size": 63488 00:24:39.806 }, 00:24:39.806 { 00:24:39.806 "name": "BaseBdev4", 00:24:39.806 "uuid": "289dece3-db1e-486d-a336-fdae863f6ff2", 00:24:39.806 "is_configured": true, 00:24:39.806 "data_offset": 2048, 00:24:39.806 "data_size": 63488 00:24:39.806 } 00:24:39.806 ] 00:24:39.806 }' 00:24:39.806 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:39.806 20:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:40.372 [2024-10-01 20:23:35.447488] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:40.372 "name": "Existed_Raid", 00:24:40.372 "uuid": "b59a4974-e22f-43fd-8503-cde7d66e7ffc", 00:24:40.372 "strip_size_kb": 64, 00:24:40.372 "state": "configuring", 00:24:40.372 "raid_level": "concat", 00:24:40.372 "superblock": true, 00:24:40.372 "num_base_bdevs": 4, 00:24:40.372 "num_base_bdevs_discovered": 2, 00:24:40.372 "num_base_bdevs_operational": 4, 00:24:40.372 "base_bdevs_list": [ 00:24:40.372 { 00:24:40.372 "name": "BaseBdev1", 00:24:40.372 "uuid": "00000000-0000-0000-0000-000000000000", 00:24:40.372 "is_configured": false, 00:24:40.372 "data_offset": 0, 00:24:40.372 "data_size": 0 00:24:40.372 }, 00:24:40.372 { 00:24:40.372 "name": null, 00:24:40.372 "uuid": "7a85b44e-5096-4d78-980b-4fa5e92fcfc4", 00:24:40.372 "is_configured": false, 00:24:40.372 "data_offset": 0, 00:24:40.372 "data_size": 63488 00:24:40.372 }, 00:24:40.372 { 00:24:40.372 "name": "BaseBdev3", 00:24:40.372 "uuid": "e1229a7c-c85d-4109-bcf3-7c232657dd40", 00:24:40.372 "is_configured": true, 00:24:40.372 "data_offset": 2048, 00:24:40.372 "data_size": 63488 00:24:40.372 }, 00:24:40.372 { 00:24:40.372 "name": "BaseBdev4", 00:24:40.372 "uuid": "289dece3-db1e-486d-a336-fdae863f6ff2", 00:24:40.372 "is_configured": true, 00:24:40.372 "data_offset": 2048, 00:24:40.372 "data_size": 63488 00:24:40.372 } 00:24:40.372 ] 00:24:40.372 }' 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:40.372 20:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:40.937 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:40.937 20:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:40.937 20:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:40.937 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:24:40.937 20:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:40.937 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:24:40.937 20:23:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:24:40.937 20:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:40.937 20:23:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:40.937 [2024-10-01 20:23:36.039397] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:24:40.937 BaseBdev1 00:24:40.937 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:40.937 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:24:40.937 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:24:40.937 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:40.937 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:24:40.937 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:40.937 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:40.937 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:40.937 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:40.937 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:40.937 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:40.937 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:24:40.937 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:40.937 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:40.937 [ 00:24:40.937 { 00:24:40.937 "name": "BaseBdev1", 00:24:40.937 "aliases": [ 00:24:40.937 "87393973-6732-4292-982c-0a59dae52329" 00:24:40.937 ], 00:24:40.937 "product_name": "Malloc disk", 00:24:40.937 "block_size": 512, 00:24:40.937 "num_blocks": 65536, 00:24:40.937 "uuid": "87393973-6732-4292-982c-0a59dae52329", 00:24:40.937 "assigned_rate_limits": { 00:24:40.937 "rw_ios_per_sec": 0, 00:24:40.937 "rw_mbytes_per_sec": 0, 00:24:40.937 "r_mbytes_per_sec": 0, 00:24:40.937 "w_mbytes_per_sec": 0 00:24:40.937 }, 00:24:40.937 "claimed": true, 00:24:40.937 "claim_type": "exclusive_write", 00:24:40.937 "zoned": false, 00:24:40.937 "supported_io_types": { 00:24:40.937 "read": true, 00:24:40.937 "write": true, 00:24:40.937 "unmap": true, 00:24:40.937 "flush": true, 00:24:40.937 "reset": true, 00:24:40.937 "nvme_admin": false, 00:24:40.937 "nvme_io": false, 00:24:40.937 "nvme_io_md": false, 00:24:40.937 "write_zeroes": true, 00:24:40.937 "zcopy": true, 00:24:40.937 "get_zone_info": false, 00:24:40.937 "zone_management": false, 00:24:40.937 "zone_append": false, 00:24:40.937 "compare": false, 00:24:40.937 "compare_and_write": false, 00:24:40.937 "abort": true, 00:24:40.937 "seek_hole": false, 00:24:40.937 "seek_data": false, 00:24:40.937 "copy": true, 00:24:40.937 "nvme_iov_md": false 00:24:40.937 }, 00:24:40.937 "memory_domains": [ 00:24:40.937 { 00:24:40.937 "dma_device_id": "system", 00:24:40.937 "dma_device_type": 1 00:24:40.937 }, 00:24:40.937 { 00:24:40.937 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:40.937 "dma_device_type": 2 00:24:40.937 } 00:24:40.937 ], 00:24:40.937 "driver_specific": {} 00:24:40.937 } 00:24:40.937 ] 00:24:40.937 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:40.937 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:40.938 "name": "Existed_Raid", 00:24:40.938 "uuid": "b59a4974-e22f-43fd-8503-cde7d66e7ffc", 00:24:40.938 "strip_size_kb": 64, 00:24:40.938 "state": "configuring", 00:24:40.938 "raid_level": "concat", 00:24:40.938 "superblock": true, 00:24:40.938 "num_base_bdevs": 4, 00:24:40.938 "num_base_bdevs_discovered": 3, 00:24:40.938 "num_base_bdevs_operational": 4, 00:24:40.938 "base_bdevs_list": [ 00:24:40.938 { 00:24:40.938 "name": "BaseBdev1", 00:24:40.938 "uuid": "87393973-6732-4292-982c-0a59dae52329", 00:24:40.938 "is_configured": true, 00:24:40.938 "data_offset": 2048, 00:24:40.938 "data_size": 63488 00:24:40.938 }, 00:24:40.938 { 00:24:40.938 "name": null, 00:24:40.938 "uuid": "7a85b44e-5096-4d78-980b-4fa5e92fcfc4", 00:24:40.938 "is_configured": false, 00:24:40.938 "data_offset": 0, 00:24:40.938 "data_size": 63488 00:24:40.938 }, 00:24:40.938 { 00:24:40.938 "name": "BaseBdev3", 00:24:40.938 "uuid": "e1229a7c-c85d-4109-bcf3-7c232657dd40", 00:24:40.938 "is_configured": true, 00:24:40.938 "data_offset": 2048, 00:24:40.938 "data_size": 63488 00:24:40.938 }, 00:24:40.938 { 00:24:40.938 "name": "BaseBdev4", 00:24:40.938 "uuid": "289dece3-db1e-486d-a336-fdae863f6ff2", 00:24:40.938 "is_configured": true, 00:24:40.938 "data_offset": 2048, 00:24:40.938 "data_size": 63488 00:24:40.938 } 00:24:40.938 ] 00:24:40.938 }' 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:40.938 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:41.505 [2024-10-01 20:23:36.619680] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:41.505 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:41.505 "name": "Existed_Raid", 00:24:41.505 "uuid": "b59a4974-e22f-43fd-8503-cde7d66e7ffc", 00:24:41.505 "strip_size_kb": 64, 00:24:41.505 "state": "configuring", 00:24:41.505 "raid_level": "concat", 00:24:41.505 "superblock": true, 00:24:41.505 "num_base_bdevs": 4, 00:24:41.505 "num_base_bdevs_discovered": 2, 00:24:41.505 "num_base_bdevs_operational": 4, 00:24:41.505 "base_bdevs_list": [ 00:24:41.505 { 00:24:41.505 "name": "BaseBdev1", 00:24:41.505 "uuid": "87393973-6732-4292-982c-0a59dae52329", 00:24:41.505 "is_configured": true, 00:24:41.505 "data_offset": 2048, 00:24:41.505 "data_size": 63488 00:24:41.505 }, 00:24:41.505 { 00:24:41.505 "name": null, 00:24:41.505 "uuid": "7a85b44e-5096-4d78-980b-4fa5e92fcfc4", 00:24:41.505 "is_configured": false, 00:24:41.505 "data_offset": 0, 00:24:41.505 "data_size": 63488 00:24:41.505 }, 00:24:41.505 { 00:24:41.505 "name": null, 00:24:41.505 "uuid": "e1229a7c-c85d-4109-bcf3-7c232657dd40", 00:24:41.505 "is_configured": false, 00:24:41.505 "data_offset": 0, 00:24:41.505 "data_size": 63488 00:24:41.505 }, 00:24:41.506 { 00:24:41.506 "name": "BaseBdev4", 00:24:41.506 "uuid": "289dece3-db1e-486d-a336-fdae863f6ff2", 00:24:41.506 "is_configured": true, 00:24:41.506 "data_offset": 2048, 00:24:41.506 "data_size": 63488 00:24:41.506 } 00:24:41.506 ] 00:24:41.506 }' 00:24:41.506 20:23:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:41.506 20:23:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:42.073 [2024-10-01 20:23:37.223897] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:42.073 "name": "Existed_Raid", 00:24:42.073 "uuid": "b59a4974-e22f-43fd-8503-cde7d66e7ffc", 00:24:42.073 "strip_size_kb": 64, 00:24:42.073 "state": "configuring", 00:24:42.073 "raid_level": "concat", 00:24:42.073 "superblock": true, 00:24:42.073 "num_base_bdevs": 4, 00:24:42.073 "num_base_bdevs_discovered": 3, 00:24:42.073 "num_base_bdevs_operational": 4, 00:24:42.073 "base_bdevs_list": [ 00:24:42.073 { 00:24:42.073 "name": "BaseBdev1", 00:24:42.073 "uuid": "87393973-6732-4292-982c-0a59dae52329", 00:24:42.073 "is_configured": true, 00:24:42.073 "data_offset": 2048, 00:24:42.073 "data_size": 63488 00:24:42.073 }, 00:24:42.073 { 00:24:42.073 "name": null, 00:24:42.073 "uuid": "7a85b44e-5096-4d78-980b-4fa5e92fcfc4", 00:24:42.073 "is_configured": false, 00:24:42.073 "data_offset": 0, 00:24:42.073 "data_size": 63488 00:24:42.073 }, 00:24:42.073 { 00:24:42.073 "name": "BaseBdev3", 00:24:42.073 "uuid": "e1229a7c-c85d-4109-bcf3-7c232657dd40", 00:24:42.073 "is_configured": true, 00:24:42.073 "data_offset": 2048, 00:24:42.073 "data_size": 63488 00:24:42.073 }, 00:24:42.073 { 00:24:42.073 "name": "BaseBdev4", 00:24:42.073 "uuid": "289dece3-db1e-486d-a336-fdae863f6ff2", 00:24:42.073 "is_configured": true, 00:24:42.073 "data_offset": 2048, 00:24:42.073 "data_size": 63488 00:24:42.073 } 00:24:42.073 ] 00:24:42.073 }' 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:42.073 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:42.640 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:24:42.640 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:42.640 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:42.640 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:42.640 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:42.640 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:24:42.640 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:24:42.640 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:42.640 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:42.640 [2024-10-01 20:23:37.836105] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:42.913 "name": "Existed_Raid", 00:24:42.913 "uuid": "b59a4974-e22f-43fd-8503-cde7d66e7ffc", 00:24:42.913 "strip_size_kb": 64, 00:24:42.913 "state": "configuring", 00:24:42.913 "raid_level": "concat", 00:24:42.913 "superblock": true, 00:24:42.913 "num_base_bdevs": 4, 00:24:42.913 "num_base_bdevs_discovered": 2, 00:24:42.913 "num_base_bdevs_operational": 4, 00:24:42.913 "base_bdevs_list": [ 00:24:42.913 { 00:24:42.913 "name": null, 00:24:42.913 "uuid": "87393973-6732-4292-982c-0a59dae52329", 00:24:42.913 "is_configured": false, 00:24:42.913 "data_offset": 0, 00:24:42.913 "data_size": 63488 00:24:42.913 }, 00:24:42.913 { 00:24:42.913 "name": null, 00:24:42.913 "uuid": "7a85b44e-5096-4d78-980b-4fa5e92fcfc4", 00:24:42.913 "is_configured": false, 00:24:42.913 "data_offset": 0, 00:24:42.913 "data_size": 63488 00:24:42.913 }, 00:24:42.913 { 00:24:42.913 "name": "BaseBdev3", 00:24:42.913 "uuid": "e1229a7c-c85d-4109-bcf3-7c232657dd40", 00:24:42.913 "is_configured": true, 00:24:42.913 "data_offset": 2048, 00:24:42.913 "data_size": 63488 00:24:42.913 }, 00:24:42.913 { 00:24:42.913 "name": "BaseBdev4", 00:24:42.913 "uuid": "289dece3-db1e-486d-a336-fdae863f6ff2", 00:24:42.913 "is_configured": true, 00:24:42.913 "data_offset": 2048, 00:24:42.913 "data_size": 63488 00:24:42.913 } 00:24:42.913 ] 00:24:42.913 }' 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:42.913 20:23:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:43.484 [2024-10-01 20:23:38.480696] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:43.484 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:43.484 "name": "Existed_Raid", 00:24:43.484 "uuid": "b59a4974-e22f-43fd-8503-cde7d66e7ffc", 00:24:43.484 "strip_size_kb": 64, 00:24:43.484 "state": "configuring", 00:24:43.484 "raid_level": "concat", 00:24:43.484 "superblock": true, 00:24:43.484 "num_base_bdevs": 4, 00:24:43.484 "num_base_bdevs_discovered": 3, 00:24:43.484 "num_base_bdevs_operational": 4, 00:24:43.484 "base_bdevs_list": [ 00:24:43.484 { 00:24:43.484 "name": null, 00:24:43.484 "uuid": "87393973-6732-4292-982c-0a59dae52329", 00:24:43.484 "is_configured": false, 00:24:43.484 "data_offset": 0, 00:24:43.484 "data_size": 63488 00:24:43.484 }, 00:24:43.484 { 00:24:43.484 "name": "BaseBdev2", 00:24:43.484 "uuid": "7a85b44e-5096-4d78-980b-4fa5e92fcfc4", 00:24:43.484 "is_configured": true, 00:24:43.484 "data_offset": 2048, 00:24:43.484 "data_size": 63488 00:24:43.484 }, 00:24:43.484 { 00:24:43.484 "name": "BaseBdev3", 00:24:43.484 "uuid": "e1229a7c-c85d-4109-bcf3-7c232657dd40", 00:24:43.484 "is_configured": true, 00:24:43.484 "data_offset": 2048, 00:24:43.484 "data_size": 63488 00:24:43.484 }, 00:24:43.484 { 00:24:43.484 "name": "BaseBdev4", 00:24:43.484 "uuid": "289dece3-db1e-486d-a336-fdae863f6ff2", 00:24:43.484 "is_configured": true, 00:24:43.484 "data_offset": 2048, 00:24:43.484 "data_size": 63488 00:24:43.484 } 00:24:43.485 ] 00:24:43.485 }' 00:24:43.485 20:23:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:43.485 20:23:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:44.051 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:24:44.051 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 87393973-6732-4292-982c-0a59dae52329 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:44.052 [2024-10-01 20:23:39.167315] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:24:44.052 NewBaseBdev 00:24:44.052 [2024-10-01 20:23:39.167930] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:24:44.052 [2024-10-01 20:23:39.167956] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:24:44.052 [2024-10-01 20:23:39.168322] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:24:44.052 [2024-10-01 20:23:39.168488] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:24:44.052 [2024-10-01 20:23:39.168510] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:24:44.052 [2024-10-01 20:23:39.168674] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:44.052 [ 00:24:44.052 { 00:24:44.052 "name": "NewBaseBdev", 00:24:44.052 "aliases": [ 00:24:44.052 "87393973-6732-4292-982c-0a59dae52329" 00:24:44.052 ], 00:24:44.052 "product_name": "Malloc disk", 00:24:44.052 "block_size": 512, 00:24:44.052 "num_blocks": 65536, 00:24:44.052 "uuid": "87393973-6732-4292-982c-0a59dae52329", 00:24:44.052 "assigned_rate_limits": { 00:24:44.052 "rw_ios_per_sec": 0, 00:24:44.052 "rw_mbytes_per_sec": 0, 00:24:44.052 "r_mbytes_per_sec": 0, 00:24:44.052 "w_mbytes_per_sec": 0 00:24:44.052 }, 00:24:44.052 "claimed": true, 00:24:44.052 "claim_type": "exclusive_write", 00:24:44.052 "zoned": false, 00:24:44.052 "supported_io_types": { 00:24:44.052 "read": true, 00:24:44.052 "write": true, 00:24:44.052 "unmap": true, 00:24:44.052 "flush": true, 00:24:44.052 "reset": true, 00:24:44.052 "nvme_admin": false, 00:24:44.052 "nvme_io": false, 00:24:44.052 "nvme_io_md": false, 00:24:44.052 "write_zeroes": true, 00:24:44.052 "zcopy": true, 00:24:44.052 "get_zone_info": false, 00:24:44.052 "zone_management": false, 00:24:44.052 "zone_append": false, 00:24:44.052 "compare": false, 00:24:44.052 "compare_and_write": false, 00:24:44.052 "abort": true, 00:24:44.052 "seek_hole": false, 00:24:44.052 "seek_data": false, 00:24:44.052 "copy": true, 00:24:44.052 "nvme_iov_md": false 00:24:44.052 }, 00:24:44.052 "memory_domains": [ 00:24:44.052 { 00:24:44.052 "dma_device_id": "system", 00:24:44.052 "dma_device_type": 1 00:24:44.052 }, 00:24:44.052 { 00:24:44.052 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:44.052 "dma_device_type": 2 00:24:44.052 } 00:24:44.052 ], 00:24:44.052 "driver_specific": {} 00:24:44.052 } 00:24:44.052 ] 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:44.052 "name": "Existed_Raid", 00:24:44.052 "uuid": "b59a4974-e22f-43fd-8503-cde7d66e7ffc", 00:24:44.052 "strip_size_kb": 64, 00:24:44.052 "state": "online", 00:24:44.052 "raid_level": "concat", 00:24:44.052 "superblock": true, 00:24:44.052 "num_base_bdevs": 4, 00:24:44.052 "num_base_bdevs_discovered": 4, 00:24:44.052 "num_base_bdevs_operational": 4, 00:24:44.052 "base_bdevs_list": [ 00:24:44.052 { 00:24:44.052 "name": "NewBaseBdev", 00:24:44.052 "uuid": "87393973-6732-4292-982c-0a59dae52329", 00:24:44.052 "is_configured": true, 00:24:44.052 "data_offset": 2048, 00:24:44.052 "data_size": 63488 00:24:44.052 }, 00:24:44.052 { 00:24:44.052 "name": "BaseBdev2", 00:24:44.052 "uuid": "7a85b44e-5096-4d78-980b-4fa5e92fcfc4", 00:24:44.052 "is_configured": true, 00:24:44.052 "data_offset": 2048, 00:24:44.052 "data_size": 63488 00:24:44.052 }, 00:24:44.052 { 00:24:44.052 "name": "BaseBdev3", 00:24:44.052 "uuid": "e1229a7c-c85d-4109-bcf3-7c232657dd40", 00:24:44.052 "is_configured": true, 00:24:44.052 "data_offset": 2048, 00:24:44.052 "data_size": 63488 00:24:44.052 }, 00:24:44.052 { 00:24:44.052 "name": "BaseBdev4", 00:24:44.052 "uuid": "289dece3-db1e-486d-a336-fdae863f6ff2", 00:24:44.052 "is_configured": true, 00:24:44.052 "data_offset": 2048, 00:24:44.052 "data_size": 63488 00:24:44.052 } 00:24:44.052 ] 00:24:44.052 }' 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:44.052 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:44.620 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:24:44.620 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:24:44.620 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:24:44.620 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:24:44.620 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:24:44.620 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:24:44.620 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:24:44.620 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:24:44.620 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:44.620 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:44.620 [2024-10-01 20:23:39.720068] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:24:44.620 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:44.620 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:44.620 "name": "Existed_Raid", 00:24:44.620 "aliases": [ 00:24:44.620 "b59a4974-e22f-43fd-8503-cde7d66e7ffc" 00:24:44.620 ], 00:24:44.620 "product_name": "Raid Volume", 00:24:44.620 "block_size": 512, 00:24:44.620 "num_blocks": 253952, 00:24:44.620 "uuid": "b59a4974-e22f-43fd-8503-cde7d66e7ffc", 00:24:44.620 "assigned_rate_limits": { 00:24:44.620 "rw_ios_per_sec": 0, 00:24:44.620 "rw_mbytes_per_sec": 0, 00:24:44.620 "r_mbytes_per_sec": 0, 00:24:44.620 "w_mbytes_per_sec": 0 00:24:44.620 }, 00:24:44.620 "claimed": false, 00:24:44.620 "zoned": false, 00:24:44.620 "supported_io_types": { 00:24:44.620 "read": true, 00:24:44.620 "write": true, 00:24:44.620 "unmap": true, 00:24:44.620 "flush": true, 00:24:44.620 "reset": true, 00:24:44.620 "nvme_admin": false, 00:24:44.620 "nvme_io": false, 00:24:44.620 "nvme_io_md": false, 00:24:44.620 "write_zeroes": true, 00:24:44.620 "zcopy": false, 00:24:44.621 "get_zone_info": false, 00:24:44.621 "zone_management": false, 00:24:44.621 "zone_append": false, 00:24:44.621 "compare": false, 00:24:44.621 "compare_and_write": false, 00:24:44.621 "abort": false, 00:24:44.621 "seek_hole": false, 00:24:44.621 "seek_data": false, 00:24:44.621 "copy": false, 00:24:44.621 "nvme_iov_md": false 00:24:44.621 }, 00:24:44.621 "memory_domains": [ 00:24:44.621 { 00:24:44.621 "dma_device_id": "system", 00:24:44.621 "dma_device_type": 1 00:24:44.621 }, 00:24:44.621 { 00:24:44.621 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:44.621 "dma_device_type": 2 00:24:44.621 }, 00:24:44.621 { 00:24:44.621 "dma_device_id": "system", 00:24:44.621 "dma_device_type": 1 00:24:44.621 }, 00:24:44.621 { 00:24:44.621 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:44.621 "dma_device_type": 2 00:24:44.621 }, 00:24:44.621 { 00:24:44.621 "dma_device_id": "system", 00:24:44.621 "dma_device_type": 1 00:24:44.621 }, 00:24:44.621 { 00:24:44.621 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:44.621 "dma_device_type": 2 00:24:44.621 }, 00:24:44.621 { 00:24:44.621 "dma_device_id": "system", 00:24:44.621 "dma_device_type": 1 00:24:44.621 }, 00:24:44.621 { 00:24:44.621 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:44.621 "dma_device_type": 2 00:24:44.621 } 00:24:44.621 ], 00:24:44.621 "driver_specific": { 00:24:44.621 "raid": { 00:24:44.621 "uuid": "b59a4974-e22f-43fd-8503-cde7d66e7ffc", 00:24:44.621 "strip_size_kb": 64, 00:24:44.621 "state": "online", 00:24:44.621 "raid_level": "concat", 00:24:44.621 "superblock": true, 00:24:44.621 "num_base_bdevs": 4, 00:24:44.621 "num_base_bdevs_discovered": 4, 00:24:44.621 "num_base_bdevs_operational": 4, 00:24:44.621 "base_bdevs_list": [ 00:24:44.621 { 00:24:44.621 "name": "NewBaseBdev", 00:24:44.621 "uuid": "87393973-6732-4292-982c-0a59dae52329", 00:24:44.621 "is_configured": true, 00:24:44.621 "data_offset": 2048, 00:24:44.621 "data_size": 63488 00:24:44.621 }, 00:24:44.621 { 00:24:44.621 "name": "BaseBdev2", 00:24:44.621 "uuid": "7a85b44e-5096-4d78-980b-4fa5e92fcfc4", 00:24:44.621 "is_configured": true, 00:24:44.621 "data_offset": 2048, 00:24:44.621 "data_size": 63488 00:24:44.621 }, 00:24:44.621 { 00:24:44.621 "name": "BaseBdev3", 00:24:44.621 "uuid": "e1229a7c-c85d-4109-bcf3-7c232657dd40", 00:24:44.621 "is_configured": true, 00:24:44.621 "data_offset": 2048, 00:24:44.621 "data_size": 63488 00:24:44.621 }, 00:24:44.621 { 00:24:44.621 "name": "BaseBdev4", 00:24:44.621 "uuid": "289dece3-db1e-486d-a336-fdae863f6ff2", 00:24:44.621 "is_configured": true, 00:24:44.621 "data_offset": 2048, 00:24:44.621 "data_size": 63488 00:24:44.621 } 00:24:44.621 ] 00:24:44.621 } 00:24:44.621 } 00:24:44.621 }' 00:24:44.621 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:24:44.621 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:24:44.621 BaseBdev2 00:24:44.621 BaseBdev3 00:24:44.621 BaseBdev4' 00:24:44.621 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:44.880 20:23:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:44.880 [2024-10-01 20:23:40.107633] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:24:44.880 [2024-10-01 20:23:40.107839] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:24:44.880 [2024-10-01 20:23:40.107974] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:24:44.880 [2024-10-01 20:23:40.108071] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:24:44.880 [2024-10-01 20:23:40.108090] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 72753 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 72753 ']' 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 72753 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:44.880 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72753 00:24:45.139 killing process with pid 72753 00:24:45.139 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:24:45.139 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:24:45.139 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72753' 00:24:45.139 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 72753 00:24:45.139 [2024-10-01 20:23:40.145951] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:24:45.139 20:23:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 72753 00:24:45.398 [2024-10-01 20:23:40.504282] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:24:47.303 20:23:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:24:47.303 00:24:47.303 real 0m13.699s 00:24:47.303 user 0m22.076s 00:24:47.303 sys 0m1.906s 00:24:47.303 20:23:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:24:47.303 20:23:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:24:47.303 ************************************ 00:24:47.303 END TEST raid_state_function_test_sb 00:24:47.303 ************************************ 00:24:47.303 20:23:42 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 4 00:24:47.303 20:23:42 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:24:47.303 20:23:42 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:24:47.303 20:23:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:24:47.303 ************************************ 00:24:47.303 START TEST raid_superblock_test 00:24:47.303 ************************************ 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test concat 4 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=73440 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 73440 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 73440 ']' 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:47.303 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:47.303 20:23:42 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:47.303 [2024-10-01 20:23:42.493137] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:24:47.303 [2024-10-01 20:23:42.493541] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73440 ] 00:24:47.562 [2024-10-01 20:23:42.658952] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:47.820 [2024-10-01 20:23:42.899255] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:24:48.079 [2024-10-01 20:23:43.107012] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:48.079 [2024-10-01 20:23:43.107085] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:48.349 malloc1 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:48.349 [2024-10-01 20:23:43.589906] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:24:48.349 [2024-10-01 20:23:43.590154] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:48.349 [2024-10-01 20:23:43.590237] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:24:48.349 [2024-10-01 20:23:43.590370] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:48.349 [2024-10-01 20:23:43.593498] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:48.349 pt1 00:24:48.349 [2024-10-01 20:23:43.593752] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:48.349 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:48.608 malloc2 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:48.608 [2024-10-01 20:23:43.650916] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:24:48.608 [2024-10-01 20:23:43.651125] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:48.608 [2024-10-01 20:23:43.651221] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:24:48.608 [2024-10-01 20:23:43.651329] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:48.608 [2024-10-01 20:23:43.654430] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:48.608 pt2 00:24:48.608 [2024-10-01 20:23:43.654607] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:48.608 malloc3 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:48.608 [2024-10-01 20:23:43.708429] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:24:48.608 [2024-10-01 20:23:43.708648] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:48.608 [2024-10-01 20:23:43.708753] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:24:48.608 [2024-10-01 20:23:43.708877] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:48.608 [2024-10-01 20:23:43.711912] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:48.608 [2024-10-01 20:23:43.712061] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:24:48.608 pt3 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:48.608 malloc4 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:48.608 [2024-10-01 20:23:43.762531] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:24:48.608 [2024-10-01 20:23:43.762804] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:48.608 [2024-10-01 20:23:43.762882] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:24:48.608 [2024-10-01 20:23:43.762991] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:48.608 [2024-10-01 20:23:43.765965] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:48.608 [2024-10-01 20:23:43.766121] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:24:48.608 pt4 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:48.608 [2024-10-01 20:23:43.770601] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:24:48.608 [2024-10-01 20:23:43.773325] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:24:48.608 [2024-10-01 20:23:43.773543] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:24:48.608 [2024-10-01 20:23:43.773651] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:24:48.608 [2024-10-01 20:23:43.773942] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:24:48.608 [2024-10-01 20:23:43.773969] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:24:48.608 [2024-10-01 20:23:43.774288] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:24:48.608 [2024-10-01 20:23:43.774510] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:24:48.608 [2024-10-01 20:23:43.774530] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:24:48.608 [2024-10-01 20:23:43.774780] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:48.608 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:48.609 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:48.609 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:48.609 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:48.609 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:48.609 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:48.609 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:48.609 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:48.609 "name": "raid_bdev1", 00:24:48.609 "uuid": "1e87235f-c108-4903-a435-226a460357f5", 00:24:48.609 "strip_size_kb": 64, 00:24:48.609 "state": "online", 00:24:48.609 "raid_level": "concat", 00:24:48.609 "superblock": true, 00:24:48.609 "num_base_bdevs": 4, 00:24:48.609 "num_base_bdevs_discovered": 4, 00:24:48.609 "num_base_bdevs_operational": 4, 00:24:48.609 "base_bdevs_list": [ 00:24:48.609 { 00:24:48.609 "name": "pt1", 00:24:48.609 "uuid": "00000000-0000-0000-0000-000000000001", 00:24:48.609 "is_configured": true, 00:24:48.609 "data_offset": 2048, 00:24:48.609 "data_size": 63488 00:24:48.609 }, 00:24:48.609 { 00:24:48.609 "name": "pt2", 00:24:48.609 "uuid": "00000000-0000-0000-0000-000000000002", 00:24:48.609 "is_configured": true, 00:24:48.609 "data_offset": 2048, 00:24:48.609 "data_size": 63488 00:24:48.609 }, 00:24:48.609 { 00:24:48.609 "name": "pt3", 00:24:48.609 "uuid": "00000000-0000-0000-0000-000000000003", 00:24:48.609 "is_configured": true, 00:24:48.609 "data_offset": 2048, 00:24:48.609 "data_size": 63488 00:24:48.609 }, 00:24:48.609 { 00:24:48.609 "name": "pt4", 00:24:48.609 "uuid": "00000000-0000-0000-0000-000000000004", 00:24:48.609 "is_configured": true, 00:24:48.609 "data_offset": 2048, 00:24:48.609 "data_size": 63488 00:24:48.609 } 00:24:48.609 ] 00:24:48.609 }' 00:24:48.609 20:23:43 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:48.609 20:23:43 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.175 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:24:49.175 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:24:49.175 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:24:49.175 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:24:49.175 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:24:49.175 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:24:49.175 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:24:49.175 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:24:49.175 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.175 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.175 [2024-10-01 20:23:44.307325] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:24:49.175 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:49.175 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:49.175 "name": "raid_bdev1", 00:24:49.175 "aliases": [ 00:24:49.175 "1e87235f-c108-4903-a435-226a460357f5" 00:24:49.175 ], 00:24:49.175 "product_name": "Raid Volume", 00:24:49.175 "block_size": 512, 00:24:49.175 "num_blocks": 253952, 00:24:49.175 "uuid": "1e87235f-c108-4903-a435-226a460357f5", 00:24:49.175 "assigned_rate_limits": { 00:24:49.175 "rw_ios_per_sec": 0, 00:24:49.175 "rw_mbytes_per_sec": 0, 00:24:49.175 "r_mbytes_per_sec": 0, 00:24:49.175 "w_mbytes_per_sec": 0 00:24:49.175 }, 00:24:49.175 "claimed": false, 00:24:49.175 "zoned": false, 00:24:49.175 "supported_io_types": { 00:24:49.175 "read": true, 00:24:49.175 "write": true, 00:24:49.175 "unmap": true, 00:24:49.175 "flush": true, 00:24:49.175 "reset": true, 00:24:49.175 "nvme_admin": false, 00:24:49.175 "nvme_io": false, 00:24:49.175 "nvme_io_md": false, 00:24:49.175 "write_zeroes": true, 00:24:49.175 "zcopy": false, 00:24:49.175 "get_zone_info": false, 00:24:49.175 "zone_management": false, 00:24:49.175 "zone_append": false, 00:24:49.175 "compare": false, 00:24:49.175 "compare_and_write": false, 00:24:49.175 "abort": false, 00:24:49.175 "seek_hole": false, 00:24:49.175 "seek_data": false, 00:24:49.175 "copy": false, 00:24:49.175 "nvme_iov_md": false 00:24:49.175 }, 00:24:49.175 "memory_domains": [ 00:24:49.175 { 00:24:49.175 "dma_device_id": "system", 00:24:49.175 "dma_device_type": 1 00:24:49.175 }, 00:24:49.175 { 00:24:49.175 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:49.175 "dma_device_type": 2 00:24:49.175 }, 00:24:49.175 { 00:24:49.175 "dma_device_id": "system", 00:24:49.175 "dma_device_type": 1 00:24:49.175 }, 00:24:49.175 { 00:24:49.175 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:49.175 "dma_device_type": 2 00:24:49.175 }, 00:24:49.175 { 00:24:49.175 "dma_device_id": "system", 00:24:49.175 "dma_device_type": 1 00:24:49.175 }, 00:24:49.175 { 00:24:49.175 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:49.175 "dma_device_type": 2 00:24:49.175 }, 00:24:49.175 { 00:24:49.175 "dma_device_id": "system", 00:24:49.175 "dma_device_type": 1 00:24:49.175 }, 00:24:49.175 { 00:24:49.175 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:49.175 "dma_device_type": 2 00:24:49.175 } 00:24:49.175 ], 00:24:49.175 "driver_specific": { 00:24:49.175 "raid": { 00:24:49.175 "uuid": "1e87235f-c108-4903-a435-226a460357f5", 00:24:49.175 "strip_size_kb": 64, 00:24:49.175 "state": "online", 00:24:49.175 "raid_level": "concat", 00:24:49.175 "superblock": true, 00:24:49.175 "num_base_bdevs": 4, 00:24:49.175 "num_base_bdevs_discovered": 4, 00:24:49.175 "num_base_bdevs_operational": 4, 00:24:49.175 "base_bdevs_list": [ 00:24:49.175 { 00:24:49.175 "name": "pt1", 00:24:49.175 "uuid": "00000000-0000-0000-0000-000000000001", 00:24:49.175 "is_configured": true, 00:24:49.175 "data_offset": 2048, 00:24:49.175 "data_size": 63488 00:24:49.175 }, 00:24:49.175 { 00:24:49.175 "name": "pt2", 00:24:49.175 "uuid": "00000000-0000-0000-0000-000000000002", 00:24:49.175 "is_configured": true, 00:24:49.175 "data_offset": 2048, 00:24:49.175 "data_size": 63488 00:24:49.175 }, 00:24:49.175 { 00:24:49.175 "name": "pt3", 00:24:49.175 "uuid": "00000000-0000-0000-0000-000000000003", 00:24:49.175 "is_configured": true, 00:24:49.175 "data_offset": 2048, 00:24:49.175 "data_size": 63488 00:24:49.175 }, 00:24:49.175 { 00:24:49.175 "name": "pt4", 00:24:49.175 "uuid": "00000000-0000-0000-0000-000000000004", 00:24:49.175 "is_configured": true, 00:24:49.175 "data_offset": 2048, 00:24:49.175 "data_size": 63488 00:24:49.175 } 00:24:49.175 ] 00:24:49.175 } 00:24:49.175 } 00:24:49.175 }' 00:24:49.175 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:24:49.175 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:24:49.175 pt2 00:24:49.175 pt3 00:24:49.175 pt4' 00:24:49.175 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.435 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.435 [2024-10-01 20:23:44.687341] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=1e87235f-c108-4903-a435-226a460357f5 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 1e87235f-c108-4903-a435-226a460357f5 ']' 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.695 [2024-10-01 20:23:44.739005] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:24:49.695 [2024-10-01 20:23:44.739050] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:24:49.695 [2024-10-01 20:23:44.739148] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:24:49.695 [2024-10-01 20:23:44.739253] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:24:49.695 [2024-10-01 20:23:44.739297] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.695 [2024-10-01 20:23:44.895092] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:24:49.695 [2024-10-01 20:23:44.897769] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:24:49.695 [2024-10-01 20:23:44.897872] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:24:49.695 [2024-10-01 20:23:44.897931] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:24:49.695 [2024-10-01 20:23:44.898007] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:24:49.695 [2024-10-01 20:23:44.898077] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:24:49.695 [2024-10-01 20:23:44.898120] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:24:49.695 [2024-10-01 20:23:44.898153] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:24:49.695 [2024-10-01 20:23:44.898175] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:24:49.695 [2024-10-01 20:23:44.898191] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:24:49.695 request: 00:24:49.695 { 00:24:49.695 "name": "raid_bdev1", 00:24:49.695 "raid_level": "concat", 00:24:49.695 "base_bdevs": [ 00:24:49.695 "malloc1", 00:24:49.695 "malloc2", 00:24:49.695 "malloc3", 00:24:49.695 "malloc4" 00:24:49.695 ], 00:24:49.695 "strip_size_kb": 64, 00:24:49.695 "superblock": false, 00:24:49.695 "method": "bdev_raid_create", 00:24:49.695 "req_id": 1 00:24:49.695 } 00:24:49.695 Got JSON-RPC error response 00:24:49.695 response: 00:24:49.695 { 00:24:49.695 "code": -17, 00:24:49.695 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:24:49.695 } 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:24:49.695 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:49.954 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:24:49.954 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:24:49.954 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:24:49.954 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.955 [2024-10-01 20:23:44.959074] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:24:49.955 [2024-10-01 20:23:44.959144] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:49.955 [2024-10-01 20:23:44.959177] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:24:49.955 [2024-10-01 20:23:44.959197] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:49.955 [2024-10-01 20:23:44.962169] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:49.955 [2024-10-01 20:23:44.962218] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:24:49.955 [2024-10-01 20:23:44.962320] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:24:49.955 [2024-10-01 20:23:44.962402] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:24:49.955 pt1 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:49.955 20:23:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:49.955 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:49.955 "name": "raid_bdev1", 00:24:49.955 "uuid": "1e87235f-c108-4903-a435-226a460357f5", 00:24:49.955 "strip_size_kb": 64, 00:24:49.955 "state": "configuring", 00:24:49.955 "raid_level": "concat", 00:24:49.955 "superblock": true, 00:24:49.955 "num_base_bdevs": 4, 00:24:49.955 "num_base_bdevs_discovered": 1, 00:24:49.955 "num_base_bdevs_operational": 4, 00:24:49.955 "base_bdevs_list": [ 00:24:49.955 { 00:24:49.955 "name": "pt1", 00:24:49.955 "uuid": "00000000-0000-0000-0000-000000000001", 00:24:49.955 "is_configured": true, 00:24:49.955 "data_offset": 2048, 00:24:49.955 "data_size": 63488 00:24:49.955 }, 00:24:49.955 { 00:24:49.955 "name": null, 00:24:49.955 "uuid": "00000000-0000-0000-0000-000000000002", 00:24:49.955 "is_configured": false, 00:24:49.955 "data_offset": 2048, 00:24:49.955 "data_size": 63488 00:24:49.955 }, 00:24:49.955 { 00:24:49.955 "name": null, 00:24:49.955 "uuid": "00000000-0000-0000-0000-000000000003", 00:24:49.955 "is_configured": false, 00:24:49.955 "data_offset": 2048, 00:24:49.955 "data_size": 63488 00:24:49.955 }, 00:24:49.955 { 00:24:49.955 "name": null, 00:24:49.955 "uuid": "00000000-0000-0000-0000-000000000004", 00:24:49.955 "is_configured": false, 00:24:49.955 "data_offset": 2048, 00:24:49.955 "data_size": 63488 00:24:49.955 } 00:24:49.955 ] 00:24:49.955 }' 00:24:49.955 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:49.955 20:23:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:50.522 [2024-10-01 20:23:45.491308] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:24:50.522 [2024-10-01 20:23:45.491408] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:50.522 [2024-10-01 20:23:45.491441] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:24:50.522 [2024-10-01 20:23:45.491461] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:50.522 [2024-10-01 20:23:45.492165] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:50.522 [2024-10-01 20:23:45.492201] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:24:50.522 [2024-10-01 20:23:45.492302] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:24:50.522 [2024-10-01 20:23:45.492373] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:24:50.522 pt2 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:50.522 [2024-10-01 20:23:45.499328] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:50.522 "name": "raid_bdev1", 00:24:50.522 "uuid": "1e87235f-c108-4903-a435-226a460357f5", 00:24:50.522 "strip_size_kb": 64, 00:24:50.522 "state": "configuring", 00:24:50.522 "raid_level": "concat", 00:24:50.522 "superblock": true, 00:24:50.522 "num_base_bdevs": 4, 00:24:50.522 "num_base_bdevs_discovered": 1, 00:24:50.522 "num_base_bdevs_operational": 4, 00:24:50.522 "base_bdevs_list": [ 00:24:50.522 { 00:24:50.522 "name": "pt1", 00:24:50.522 "uuid": "00000000-0000-0000-0000-000000000001", 00:24:50.522 "is_configured": true, 00:24:50.522 "data_offset": 2048, 00:24:50.522 "data_size": 63488 00:24:50.522 }, 00:24:50.522 { 00:24:50.522 "name": null, 00:24:50.522 "uuid": "00000000-0000-0000-0000-000000000002", 00:24:50.522 "is_configured": false, 00:24:50.522 "data_offset": 0, 00:24:50.522 "data_size": 63488 00:24:50.522 }, 00:24:50.522 { 00:24:50.522 "name": null, 00:24:50.522 "uuid": "00000000-0000-0000-0000-000000000003", 00:24:50.522 "is_configured": false, 00:24:50.522 "data_offset": 2048, 00:24:50.522 "data_size": 63488 00:24:50.522 }, 00:24:50.522 { 00:24:50.522 "name": null, 00:24:50.522 "uuid": "00000000-0000-0000-0000-000000000004", 00:24:50.522 "is_configured": false, 00:24:50.522 "data_offset": 2048, 00:24:50.522 "data_size": 63488 00:24:50.522 } 00:24:50.522 ] 00:24:50.522 }' 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:50.522 20:23:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:50.780 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:24:50.781 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:24:50.781 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:24:50.781 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:50.781 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:50.781 [2024-10-01 20:23:46.031515] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:24:50.781 [2024-10-01 20:23:46.031596] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:50.781 [2024-10-01 20:23:46.031633] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:24:50.781 [2024-10-01 20:23:46.031650] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:50.781 [2024-10-01 20:23:46.032290] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:50.781 [2024-10-01 20:23:46.032317] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:24:50.781 [2024-10-01 20:23:46.032436] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:24:50.781 [2024-10-01 20:23:46.032470] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:24:51.040 pt2 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:51.040 [2024-10-01 20:23:46.039452] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:24:51.040 [2024-10-01 20:23:46.039507] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:51.040 [2024-10-01 20:23:46.039536] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:24:51.040 [2024-10-01 20:23:46.039551] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:51.040 [2024-10-01 20:23:46.040043] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:51.040 [2024-10-01 20:23:46.040068] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:24:51.040 [2024-10-01 20:23:46.040151] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:24:51.040 [2024-10-01 20:23:46.040189] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:24:51.040 pt3 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:51.040 [2024-10-01 20:23:46.047422] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:24:51.040 [2024-10-01 20:23:46.047501] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:51.040 [2024-10-01 20:23:46.047531] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:24:51.040 [2024-10-01 20:23:46.047545] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:51.040 [2024-10-01 20:23:46.048040] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:51.040 [2024-10-01 20:23:46.048072] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:24:51.040 [2024-10-01 20:23:46.048153] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:24:51.040 [2024-10-01 20:23:46.048181] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:24:51.040 [2024-10-01 20:23:46.048350] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:24:51.040 [2024-10-01 20:23:46.048366] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:24:51.040 [2024-10-01 20:23:46.048670] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:24:51.040 [2024-10-01 20:23:46.048884] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:24:51.040 [2024-10-01 20:23:46.048913] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:24:51.040 [2024-10-01 20:23:46.049076] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:51.040 pt4 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:51.040 "name": "raid_bdev1", 00:24:51.040 "uuid": "1e87235f-c108-4903-a435-226a460357f5", 00:24:51.040 "strip_size_kb": 64, 00:24:51.040 "state": "online", 00:24:51.040 "raid_level": "concat", 00:24:51.040 "superblock": true, 00:24:51.040 "num_base_bdevs": 4, 00:24:51.040 "num_base_bdevs_discovered": 4, 00:24:51.040 "num_base_bdevs_operational": 4, 00:24:51.040 "base_bdevs_list": [ 00:24:51.040 { 00:24:51.040 "name": "pt1", 00:24:51.040 "uuid": "00000000-0000-0000-0000-000000000001", 00:24:51.040 "is_configured": true, 00:24:51.040 "data_offset": 2048, 00:24:51.040 "data_size": 63488 00:24:51.040 }, 00:24:51.040 { 00:24:51.040 "name": "pt2", 00:24:51.040 "uuid": "00000000-0000-0000-0000-000000000002", 00:24:51.040 "is_configured": true, 00:24:51.040 "data_offset": 2048, 00:24:51.040 "data_size": 63488 00:24:51.040 }, 00:24:51.040 { 00:24:51.040 "name": "pt3", 00:24:51.040 "uuid": "00000000-0000-0000-0000-000000000003", 00:24:51.040 "is_configured": true, 00:24:51.040 "data_offset": 2048, 00:24:51.040 "data_size": 63488 00:24:51.040 }, 00:24:51.040 { 00:24:51.040 "name": "pt4", 00:24:51.040 "uuid": "00000000-0000-0000-0000-000000000004", 00:24:51.040 "is_configured": true, 00:24:51.040 "data_offset": 2048, 00:24:51.040 "data_size": 63488 00:24:51.040 } 00:24:51.040 ] 00:24:51.040 }' 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:51.040 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:51.608 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:24:51.608 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:24:51.608 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:24:51.608 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:24:51.608 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:24:51.608 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:24:51.608 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:24:51.608 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:51.608 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:51.608 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:24:51.608 [2024-10-01 20:23:46.576083] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:24:51.608 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:51.608 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:24:51.608 "name": "raid_bdev1", 00:24:51.608 "aliases": [ 00:24:51.608 "1e87235f-c108-4903-a435-226a460357f5" 00:24:51.608 ], 00:24:51.608 "product_name": "Raid Volume", 00:24:51.608 "block_size": 512, 00:24:51.608 "num_blocks": 253952, 00:24:51.608 "uuid": "1e87235f-c108-4903-a435-226a460357f5", 00:24:51.608 "assigned_rate_limits": { 00:24:51.608 "rw_ios_per_sec": 0, 00:24:51.608 "rw_mbytes_per_sec": 0, 00:24:51.608 "r_mbytes_per_sec": 0, 00:24:51.608 "w_mbytes_per_sec": 0 00:24:51.608 }, 00:24:51.608 "claimed": false, 00:24:51.608 "zoned": false, 00:24:51.608 "supported_io_types": { 00:24:51.608 "read": true, 00:24:51.608 "write": true, 00:24:51.608 "unmap": true, 00:24:51.608 "flush": true, 00:24:51.608 "reset": true, 00:24:51.608 "nvme_admin": false, 00:24:51.608 "nvme_io": false, 00:24:51.608 "nvme_io_md": false, 00:24:51.608 "write_zeroes": true, 00:24:51.608 "zcopy": false, 00:24:51.608 "get_zone_info": false, 00:24:51.608 "zone_management": false, 00:24:51.608 "zone_append": false, 00:24:51.608 "compare": false, 00:24:51.608 "compare_and_write": false, 00:24:51.608 "abort": false, 00:24:51.608 "seek_hole": false, 00:24:51.608 "seek_data": false, 00:24:51.608 "copy": false, 00:24:51.608 "nvme_iov_md": false 00:24:51.608 }, 00:24:51.608 "memory_domains": [ 00:24:51.608 { 00:24:51.608 "dma_device_id": "system", 00:24:51.608 "dma_device_type": 1 00:24:51.608 }, 00:24:51.608 { 00:24:51.608 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:51.608 "dma_device_type": 2 00:24:51.608 }, 00:24:51.608 { 00:24:51.608 "dma_device_id": "system", 00:24:51.608 "dma_device_type": 1 00:24:51.608 }, 00:24:51.608 { 00:24:51.608 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:51.608 "dma_device_type": 2 00:24:51.608 }, 00:24:51.608 { 00:24:51.608 "dma_device_id": "system", 00:24:51.608 "dma_device_type": 1 00:24:51.608 }, 00:24:51.608 { 00:24:51.608 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:51.608 "dma_device_type": 2 00:24:51.608 }, 00:24:51.608 { 00:24:51.608 "dma_device_id": "system", 00:24:51.608 "dma_device_type": 1 00:24:51.608 }, 00:24:51.608 { 00:24:51.608 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:24:51.608 "dma_device_type": 2 00:24:51.608 } 00:24:51.608 ], 00:24:51.608 "driver_specific": { 00:24:51.608 "raid": { 00:24:51.608 "uuid": "1e87235f-c108-4903-a435-226a460357f5", 00:24:51.608 "strip_size_kb": 64, 00:24:51.608 "state": "online", 00:24:51.608 "raid_level": "concat", 00:24:51.608 "superblock": true, 00:24:51.608 "num_base_bdevs": 4, 00:24:51.608 "num_base_bdevs_discovered": 4, 00:24:51.608 "num_base_bdevs_operational": 4, 00:24:51.608 "base_bdevs_list": [ 00:24:51.608 { 00:24:51.608 "name": "pt1", 00:24:51.608 "uuid": "00000000-0000-0000-0000-000000000001", 00:24:51.608 "is_configured": true, 00:24:51.608 "data_offset": 2048, 00:24:51.608 "data_size": 63488 00:24:51.608 }, 00:24:51.608 { 00:24:51.608 "name": "pt2", 00:24:51.608 "uuid": "00000000-0000-0000-0000-000000000002", 00:24:51.608 "is_configured": true, 00:24:51.608 "data_offset": 2048, 00:24:51.608 "data_size": 63488 00:24:51.608 }, 00:24:51.608 { 00:24:51.608 "name": "pt3", 00:24:51.608 "uuid": "00000000-0000-0000-0000-000000000003", 00:24:51.608 "is_configured": true, 00:24:51.608 "data_offset": 2048, 00:24:51.608 "data_size": 63488 00:24:51.608 }, 00:24:51.608 { 00:24:51.608 "name": "pt4", 00:24:51.608 "uuid": "00000000-0000-0000-0000-000000000004", 00:24:51.608 "is_configured": true, 00:24:51.608 "data_offset": 2048, 00:24:51.608 "data_size": 63488 00:24:51.608 } 00:24:51.608 ] 00:24:51.608 } 00:24:51.608 } 00:24:51.608 }' 00:24:51.608 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:24:51.608 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:24:51.608 pt2 00:24:51.608 pt3 00:24:51.609 pt4' 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:51.609 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:51.868 [2024-10-01 20:23:46.948176] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 1e87235f-c108-4903-a435-226a460357f5 '!=' 1e87235f-c108-4903-a435-226a460357f5 ']' 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 73440 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 73440 ']' 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 73440 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:24:51.868 20:23:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:51.868 20:23:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73440 00:24:51.868 20:23:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:24:51.868 killing process with pid 73440 00:24:51.868 20:23:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:24:51.868 20:23:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73440' 00:24:51.868 20:23:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 73440 00:24:51.868 [2024-10-01 20:23:47.028956] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:24:51.868 20:23:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 73440 00:24:51.868 [2024-10-01 20:23:47.029076] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:24:51.868 [2024-10-01 20:23:47.029184] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:24:51.868 [2024-10-01 20:23:47.029202] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:24:52.436 [2024-10-01 20:23:47.389205] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:24:54.386 20:23:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:24:54.386 00:24:54.386 real 0m6.844s 00:24:54.386 user 0m9.851s 00:24:54.386 sys 0m0.969s 00:24:54.386 20:23:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:24:54.386 ************************************ 00:24:54.386 END TEST raid_superblock_test 00:24:54.386 20:23:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:24:54.386 ************************************ 00:24:54.386 20:23:49 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 4 read 00:24:54.386 20:23:49 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:24:54.386 20:23:49 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:24:54.386 20:23:49 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:24:54.386 ************************************ 00:24:54.386 START TEST raid_read_error_test 00:24:54.386 ************************************ 00:24:54.386 20:23:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 4 read 00:24:54.386 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:24:54.386 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:24:54.386 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:24:54.386 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:24:54.386 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:54.386 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:24:54.386 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:24:54.386 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:54.386 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:24:54.386 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:24:54.386 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:54.386 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:24:54.386 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:24:54.386 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:54.386 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.ocZyWMsQCR 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73716 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73716 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 73716 ']' 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:54.387 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:54.387 20:23:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:54.387 [2024-10-01 20:23:49.431681] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:24:54.387 [2024-10-01 20:23:49.431874] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73716 ] 00:24:54.387 [2024-10-01 20:23:49.610038] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:54.645 [2024-10-01 20:23:49.858410] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:24:54.904 [2024-10-01 20:23:50.064363] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:54.904 [2024-10-01 20:23:50.064442] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:55.471 BaseBdev1_malloc 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:55.471 true 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:55.471 [2024-10-01 20:23:50.537838] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:24:55.471 [2024-10-01 20:23:50.538044] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:55.471 [2024-10-01 20:23:50.538118] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:24:55.471 [2024-10-01 20:23:50.538248] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:55.471 [2024-10-01 20:23:50.541152] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:55.471 [2024-10-01 20:23:50.541331] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:24:55.471 BaseBdev1 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:55.471 BaseBdev2_malloc 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:55.471 true 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:55.471 [2024-10-01 20:23:50.598818] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:24:55.471 [2024-10-01 20:23:50.599014] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:55.471 [2024-10-01 20:23:50.599085] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:24:55.471 [2024-10-01 20:23:50.599196] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:55.471 [2024-10-01 20:23:50.602111] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:55.471 [2024-10-01 20:23:50.602272] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:24:55.471 BaseBdev2 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:55.471 BaseBdev3_malloc 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:55.471 true 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.471 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:55.472 [2024-10-01 20:23:50.659539] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:24:55.472 [2024-10-01 20:23:50.659753] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:55.472 [2024-10-01 20:23:50.659827] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:24:55.472 [2024-10-01 20:23:50.659988] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:55.472 [2024-10-01 20:23:50.663010] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:55.472 [2024-10-01 20:23:50.663171] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:24:55.472 BaseBdev3 00:24:55.472 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.472 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:24:55.472 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:24:55.472 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.472 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:55.472 BaseBdev4_malloc 00:24:55.472 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.472 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:24:55.472 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.472 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:55.472 true 00:24:55.472 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.472 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:24:55.472 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.472 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:55.472 [2024-10-01 20:23:50.720535] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:24:55.472 [2024-10-01 20:23:50.720806] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:24:55.472 [2024-10-01 20:23:50.720861] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:24:55.472 [2024-10-01 20:23:50.720893] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:24:55.731 [2024-10-01 20:23:50.724695] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:24:55.731 BaseBdev4 00:24:55.731 [2024-10-01 20:23:50.724946] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:55.731 [2024-10-01 20:23:50.729281] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:24:55.731 [2024-10-01 20:23:50.731945] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:24:55.731 [2024-10-01 20:23:50.732067] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:24:55.731 [2024-10-01 20:23:50.732163] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:24:55.731 [2024-10-01 20:23:50.732477] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:24:55.731 [2024-10-01 20:23:50.732501] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:24:55.731 [2024-10-01 20:23:50.732819] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:24:55.731 [2024-10-01 20:23:50.733039] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:24:55.731 [2024-10-01 20:23:50.733071] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:24:55.731 [2024-10-01 20:23:50.733337] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:55.731 "name": "raid_bdev1", 00:24:55.731 "uuid": "20624301-6237-487a-afe6-45aae48ea66c", 00:24:55.731 "strip_size_kb": 64, 00:24:55.731 "state": "online", 00:24:55.731 "raid_level": "concat", 00:24:55.731 "superblock": true, 00:24:55.731 "num_base_bdevs": 4, 00:24:55.731 "num_base_bdevs_discovered": 4, 00:24:55.731 "num_base_bdevs_operational": 4, 00:24:55.731 "base_bdevs_list": [ 00:24:55.731 { 00:24:55.731 "name": "BaseBdev1", 00:24:55.731 "uuid": "0a362730-2f4a-55d2-956e-b66961a1123e", 00:24:55.731 "is_configured": true, 00:24:55.731 "data_offset": 2048, 00:24:55.731 "data_size": 63488 00:24:55.731 }, 00:24:55.731 { 00:24:55.731 "name": "BaseBdev2", 00:24:55.731 "uuid": "1577e08f-8a1c-52ff-89e8-13bd2ae8ea46", 00:24:55.731 "is_configured": true, 00:24:55.731 "data_offset": 2048, 00:24:55.731 "data_size": 63488 00:24:55.731 }, 00:24:55.731 { 00:24:55.731 "name": "BaseBdev3", 00:24:55.731 "uuid": "067b00ba-ccd8-5f44-b246-bf3f4bf777fd", 00:24:55.731 "is_configured": true, 00:24:55.731 "data_offset": 2048, 00:24:55.731 "data_size": 63488 00:24:55.731 }, 00:24:55.731 { 00:24:55.731 "name": "BaseBdev4", 00:24:55.731 "uuid": "2207ddd4-a596-53de-aae0-bb580f0165ae", 00:24:55.731 "is_configured": true, 00:24:55.731 "data_offset": 2048, 00:24:55.731 "data_size": 63488 00:24:55.731 } 00:24:55.731 ] 00:24:55.731 }' 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:55.731 20:23:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:56.295 20:23:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:24:56.295 20:23:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:24:56.295 [2024-10-01 20:23:51.371011] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:24:57.225 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:24:57.225 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:57.225 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:57.225 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:57.225 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:24:57.225 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:24:57.225 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:24:57.225 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:24:57.225 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:24:57.225 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:24:57.225 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:24:57.225 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:24:57.225 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:24:57.225 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:24:57.226 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:24:57.226 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:24:57.226 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:24:57.226 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:24:57.226 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:57.226 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:57.226 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:24:57.226 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:57.226 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:24:57.226 "name": "raid_bdev1", 00:24:57.226 "uuid": "20624301-6237-487a-afe6-45aae48ea66c", 00:24:57.226 "strip_size_kb": 64, 00:24:57.226 "state": "online", 00:24:57.226 "raid_level": "concat", 00:24:57.226 "superblock": true, 00:24:57.226 "num_base_bdevs": 4, 00:24:57.226 "num_base_bdevs_discovered": 4, 00:24:57.226 "num_base_bdevs_operational": 4, 00:24:57.226 "base_bdevs_list": [ 00:24:57.226 { 00:24:57.226 "name": "BaseBdev1", 00:24:57.226 "uuid": "0a362730-2f4a-55d2-956e-b66961a1123e", 00:24:57.226 "is_configured": true, 00:24:57.226 "data_offset": 2048, 00:24:57.226 "data_size": 63488 00:24:57.226 }, 00:24:57.226 { 00:24:57.226 "name": "BaseBdev2", 00:24:57.226 "uuid": "1577e08f-8a1c-52ff-89e8-13bd2ae8ea46", 00:24:57.226 "is_configured": true, 00:24:57.226 "data_offset": 2048, 00:24:57.226 "data_size": 63488 00:24:57.226 }, 00:24:57.226 { 00:24:57.226 "name": "BaseBdev3", 00:24:57.226 "uuid": "067b00ba-ccd8-5f44-b246-bf3f4bf777fd", 00:24:57.226 "is_configured": true, 00:24:57.226 "data_offset": 2048, 00:24:57.226 "data_size": 63488 00:24:57.226 }, 00:24:57.226 { 00:24:57.226 "name": "BaseBdev4", 00:24:57.226 "uuid": "2207ddd4-a596-53de-aae0-bb580f0165ae", 00:24:57.226 "is_configured": true, 00:24:57.226 "data_offset": 2048, 00:24:57.226 "data_size": 63488 00:24:57.226 } 00:24:57.226 ] 00:24:57.226 }' 00:24:57.226 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:24:57.226 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:57.790 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:24:57.790 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:57.790 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:57.790 [2024-10-01 20:23:52.794382] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:24:57.790 [2024-10-01 20:23:52.794422] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:24:57.790 { 00:24:57.790 "results": [ 00:24:57.790 { 00:24:57.790 "job": "raid_bdev1", 00:24:57.790 "core_mask": "0x1", 00:24:57.790 "workload": "randrw", 00:24:57.790 "percentage": 50, 00:24:57.790 "status": "finished", 00:24:57.790 "queue_depth": 1, 00:24:57.790 "io_size": 131072, 00:24:57.790 "runtime": 1.420605, 00:24:57.790 "iops": 10247.042633244288, 00:24:57.790 "mibps": 1280.880329155536, 00:24:57.790 "io_failed": 1, 00:24:57.791 "io_timeout": 0, 00:24:57.791 "avg_latency_us": 136.04807453571294, 00:24:57.791 "min_latency_us": 40.261818181818185, 00:24:57.791 "max_latency_us": 1921.3963636363637 00:24:57.791 } 00:24:57.791 ], 00:24:57.791 "core_count": 1 00:24:57.791 } 00:24:57.791 [2024-10-01 20:23:52.797906] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:24:57.791 [2024-10-01 20:23:52.797982] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:24:57.791 [2024-10-01 20:23:52.798045] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:24:57.791 [2024-10-01 20:23:52.798082] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:24:57.791 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:57.791 20:23:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73716 00:24:57.791 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 73716 ']' 00:24:57.791 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 73716 00:24:57.791 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:24:57.791 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:57.791 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73716 00:24:57.791 killing process with pid 73716 00:24:57.791 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:24:57.791 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:24:57.791 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73716' 00:24:57.791 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 73716 00:24:57.791 [2024-10-01 20:23:52.833009] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:24:57.791 20:23:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 73716 00:24:58.048 [2024-10-01 20:23:53.133667] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:24:59.962 20:23:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.ocZyWMsQCR 00:24:59.962 20:23:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:24:59.962 20:23:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:24:59.962 20:23:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.70 00:24:59.962 20:23:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:24:59.962 20:23:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:24:59.962 20:23:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:24:59.962 20:23:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.70 != \0\.\0\0 ]] 00:24:59.962 00:24:59.962 real 0m5.737s 00:24:59.962 user 0m6.740s 00:24:59.962 sys 0m0.744s 00:24:59.962 20:23:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:24:59.962 20:23:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:59.962 ************************************ 00:24:59.962 END TEST raid_read_error_test 00:24:59.962 ************************************ 00:24:59.962 20:23:55 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 4 write 00:24:59.962 20:23:55 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:24:59.962 20:23:55 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:24:59.962 20:23:55 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:24:59.962 ************************************ 00:24:59.962 START TEST raid_write_error_test 00:24:59.962 ************************************ 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 4 write 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:24:59.962 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:24:59.963 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:24:59.963 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:24:59.963 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.ewwcrATj4m 00:24:59.963 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73874 00:24:59.963 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:24:59.963 20:23:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73874 00:24:59.963 20:23:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 73874 ']' 00:24:59.963 20:23:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:59.963 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:59.963 20:23:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:59.963 20:23:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:59.963 20:23:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:59.963 20:23:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:24:59.963 [2024-10-01 20:23:55.212180] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:24:59.963 [2024-10-01 20:23:55.212546] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73874 ] 00:25:00.221 [2024-10-01 20:23:55.392965] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:00.479 [2024-10-01 20:23:55.659135] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:25:00.737 [2024-10-01 20:23:55.861695] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:00.737 [2024-10-01 20:23:55.862064] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:01.304 BaseBdev1_malloc 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:01.304 true 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:01.304 [2024-10-01 20:23:56.340327] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:25:01.304 [2024-10-01 20:23:56.340426] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:01.304 [2024-10-01 20:23:56.340470] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:25:01.304 [2024-10-01 20:23:56.340503] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:01.304 [2024-10-01 20:23:56.343901] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:01.304 [2024-10-01 20:23:56.343980] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:25:01.304 BaseBdev1 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:01.304 BaseBdev2_malloc 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:01.304 true 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:01.304 [2024-10-01 20:23:56.401293] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:25:01.304 [2024-10-01 20:23:56.401380] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:01.304 [2024-10-01 20:23:56.401433] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:25:01.304 [2024-10-01 20:23:56.401462] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:01.304 [2024-10-01 20:23:56.404596] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:01.304 [2024-10-01 20:23:56.404657] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:25:01.304 BaseBdev2 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:01.304 BaseBdev3_malloc 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:01.304 true 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:01.304 [2024-10-01 20:23:56.462992] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:25:01.304 [2024-10-01 20:23:56.463070] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:01.304 [2024-10-01 20:23:56.463111] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:25:01.304 [2024-10-01 20:23:56.463139] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:01.304 [2024-10-01 20:23:56.466233] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:01.304 [2024-10-01 20:23:56.466438] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:25:01.304 BaseBdev3 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:01.304 BaseBdev4_malloc 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:01.304 true 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:01.304 [2024-10-01 20:23:56.523818] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:25:01.304 [2024-10-01 20:23:56.523912] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:01.304 [2024-10-01 20:23:56.523954] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:25:01.304 [2024-10-01 20:23:56.523982] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:01.304 [2024-10-01 20:23:56.527111] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:01.304 [2024-10-01 20:23:56.527334] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:25:01.304 BaseBdev4 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:01.304 [2024-10-01 20:23:56.532190] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:25:01.304 [2024-10-01 20:23:56.534749] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:25:01.304 [2024-10-01 20:23:56.535071] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:25:01.304 [2024-10-01 20:23:56.535367] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:25:01.304 [2024-10-01 20:23:56.535960] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:25:01.304 [2024-10-01 20:23:56.536159] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:25:01.304 [2024-10-01 20:23:56.536650] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:25:01.304 [2024-10-01 20:23:56.537131] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:25:01.304 [2024-10-01 20:23:56.537312] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:25:01.304 [2024-10-01 20:23:56.537887] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:01.304 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.305 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:25:01.305 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:01.305 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:01.305 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:25:01.305 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:25:01.305 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:01.305 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:01.305 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:01.305 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:01.305 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:01.305 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:01.305 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:01.305 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:01.305 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:01.563 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:01.563 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:01.563 "name": "raid_bdev1", 00:25:01.563 "uuid": "4a344212-354e-4236-a258-6608c5515500", 00:25:01.563 "strip_size_kb": 64, 00:25:01.563 "state": "online", 00:25:01.563 "raid_level": "concat", 00:25:01.563 "superblock": true, 00:25:01.563 "num_base_bdevs": 4, 00:25:01.563 "num_base_bdevs_discovered": 4, 00:25:01.563 "num_base_bdevs_operational": 4, 00:25:01.563 "base_bdevs_list": [ 00:25:01.563 { 00:25:01.563 "name": "BaseBdev1", 00:25:01.563 "uuid": "e61fcbfc-2c8b-5a9e-95f4-afee4446eef6", 00:25:01.563 "is_configured": true, 00:25:01.563 "data_offset": 2048, 00:25:01.563 "data_size": 63488 00:25:01.563 }, 00:25:01.563 { 00:25:01.563 "name": "BaseBdev2", 00:25:01.563 "uuid": "3364ff66-0f2e-5d4d-bfb8-6c95d7da465d", 00:25:01.563 "is_configured": true, 00:25:01.563 "data_offset": 2048, 00:25:01.563 "data_size": 63488 00:25:01.563 }, 00:25:01.563 { 00:25:01.563 "name": "BaseBdev3", 00:25:01.563 "uuid": "e019cf9c-e995-5fc1-bb93-8cd70172d477", 00:25:01.563 "is_configured": true, 00:25:01.563 "data_offset": 2048, 00:25:01.563 "data_size": 63488 00:25:01.563 }, 00:25:01.563 { 00:25:01.563 "name": "BaseBdev4", 00:25:01.563 "uuid": "56299ecd-8789-54bf-aa74-4e1db5d8a63d", 00:25:01.563 "is_configured": true, 00:25:01.563 "data_offset": 2048, 00:25:01.563 "data_size": 63488 00:25:01.563 } 00:25:01.563 ] 00:25:01.563 }' 00:25:01.563 20:23:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:01.563 20:23:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:01.821 20:23:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:25:01.821 20:23:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:25:02.078 [2024-10-01 20:23:57.175366] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:03.012 "name": "raid_bdev1", 00:25:03.012 "uuid": "4a344212-354e-4236-a258-6608c5515500", 00:25:03.012 "strip_size_kb": 64, 00:25:03.012 "state": "online", 00:25:03.012 "raid_level": "concat", 00:25:03.012 "superblock": true, 00:25:03.012 "num_base_bdevs": 4, 00:25:03.012 "num_base_bdevs_discovered": 4, 00:25:03.012 "num_base_bdevs_operational": 4, 00:25:03.012 "base_bdevs_list": [ 00:25:03.012 { 00:25:03.012 "name": "BaseBdev1", 00:25:03.012 "uuid": "e61fcbfc-2c8b-5a9e-95f4-afee4446eef6", 00:25:03.012 "is_configured": true, 00:25:03.012 "data_offset": 2048, 00:25:03.012 "data_size": 63488 00:25:03.012 }, 00:25:03.012 { 00:25:03.012 "name": "BaseBdev2", 00:25:03.012 "uuid": "3364ff66-0f2e-5d4d-bfb8-6c95d7da465d", 00:25:03.012 "is_configured": true, 00:25:03.012 "data_offset": 2048, 00:25:03.012 "data_size": 63488 00:25:03.012 }, 00:25:03.012 { 00:25:03.012 "name": "BaseBdev3", 00:25:03.012 "uuid": "e019cf9c-e995-5fc1-bb93-8cd70172d477", 00:25:03.012 "is_configured": true, 00:25:03.012 "data_offset": 2048, 00:25:03.012 "data_size": 63488 00:25:03.012 }, 00:25:03.012 { 00:25:03.012 "name": "BaseBdev4", 00:25:03.012 "uuid": "56299ecd-8789-54bf-aa74-4e1db5d8a63d", 00:25:03.012 "is_configured": true, 00:25:03.012 "data_offset": 2048, 00:25:03.012 "data_size": 63488 00:25:03.012 } 00:25:03.012 ] 00:25:03.012 }' 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:03.012 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:03.579 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:25:03.579 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:03.579 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:03.579 [2024-10-01 20:23:58.574012] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:25:03.579 [2024-10-01 20:23:58.574108] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:25:03.579 [2024-10-01 20:23:58.577457] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:25:03.579 [2024-10-01 20:23:58.577688] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:03.579 [2024-10-01 20:23:58.577835] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:25:03.579 [2024-10-01 20:23:58.578015] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:25:03.579 { 00:25:03.579 "results": [ 00:25:03.579 { 00:25:03.579 "job": "raid_bdev1", 00:25:03.579 "core_mask": "0x1", 00:25:03.579 "workload": "randrw", 00:25:03.579 "percentage": 50, 00:25:03.579 "status": "finished", 00:25:03.579 "queue_depth": 1, 00:25:03.579 "io_size": 131072, 00:25:03.579 "runtime": 1.395864, 00:25:03.579 "iops": 10029.630393791946, 00:25:03.579 "mibps": 1253.7037992239932, 00:25:03.579 "io_failed": 1, 00:25:03.579 "io_timeout": 0, 00:25:03.579 "avg_latency_us": 139.2652815707969, 00:25:03.579 "min_latency_us": 41.42545454545454, 00:25:03.579 "max_latency_us": 1906.5018181818182 00:25:03.579 } 00:25:03.579 ], 00:25:03.579 "core_count": 1 00:25:03.579 } 00:25:03.579 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:03.579 20:23:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73874 00:25:03.579 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 73874 ']' 00:25:03.579 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 73874 00:25:03.580 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:25:03.580 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:25:03.580 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73874 00:25:03.580 killing process with pid 73874 00:25:03.580 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:25:03.580 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:25:03.580 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73874' 00:25:03.580 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 73874 00:25:03.580 [2024-10-01 20:23:58.612835] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:25:03.580 20:23:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 73874 00:25:03.839 [2024-10-01 20:23:58.901286] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:25:05.748 20:24:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.ewwcrATj4m 00:25:05.748 20:24:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:25:05.748 20:24:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:25:05.748 20:24:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:25:05.748 20:24:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:25:05.748 20:24:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:25:05.748 20:24:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:25:05.748 ************************************ 00:25:05.748 END TEST raid_write_error_test 00:25:05.748 ************************************ 00:25:05.748 20:24:00 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:25:05.748 00:25:05.748 real 0m5.702s 00:25:05.748 user 0m6.731s 00:25:05.748 sys 0m0.715s 00:25:05.748 20:24:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:25:05.748 20:24:00 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:05.748 20:24:00 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:25:05.748 20:24:00 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 4 false 00:25:05.748 20:24:00 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:25:05.748 20:24:00 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:25:05.748 20:24:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:25:05.748 ************************************ 00:25:05.748 START TEST raid_state_function_test 00:25:05.748 ************************************ 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 4 false 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:25:05.748 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=74026 00:25:05.749 Process raid pid: 74026 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 74026' 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 74026 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 74026 ']' 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:25:05.749 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:25:05.749 20:24:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:05.749 [2024-10-01 20:24:00.961838] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:25:05.749 [2024-10-01 20:24:00.962020] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:25:06.009 [2024-10-01 20:24:01.138787] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:06.267 [2024-10-01 20:24:01.390976] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:25:06.527 [2024-10-01 20:24:01.603051] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:06.527 [2024-10-01 20:24:01.603129] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:06.786 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:25:06.786 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:25:06.786 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:25:06.786 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:06.786 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:06.786 [2024-10-01 20:24:02.032158] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:25:06.786 [2024-10-01 20:24:02.032221] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:25:06.786 [2024-10-01 20:24:02.032239] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:25:06.786 [2024-10-01 20:24:02.032258] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:25:06.786 [2024-10-01 20:24:02.032269] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:25:06.786 [2024-10-01 20:24:02.032288] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:25:06.786 [2024-10-01 20:24:02.032299] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:25:06.786 [2024-10-01 20:24:02.032314] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:25:06.786 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:06.786 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:06.786 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:06.786 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:06.786 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:06.786 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:06.786 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:06.786 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:06.786 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:06.786 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:06.786 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:07.045 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:07.045 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:07.045 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:07.045 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:07.045 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:07.045 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:07.045 "name": "Existed_Raid", 00:25:07.045 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:07.045 "strip_size_kb": 0, 00:25:07.045 "state": "configuring", 00:25:07.045 "raid_level": "raid1", 00:25:07.045 "superblock": false, 00:25:07.045 "num_base_bdevs": 4, 00:25:07.045 "num_base_bdevs_discovered": 0, 00:25:07.045 "num_base_bdevs_operational": 4, 00:25:07.045 "base_bdevs_list": [ 00:25:07.045 { 00:25:07.045 "name": "BaseBdev1", 00:25:07.045 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:07.045 "is_configured": false, 00:25:07.045 "data_offset": 0, 00:25:07.045 "data_size": 0 00:25:07.045 }, 00:25:07.045 { 00:25:07.045 "name": "BaseBdev2", 00:25:07.045 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:07.045 "is_configured": false, 00:25:07.045 "data_offset": 0, 00:25:07.045 "data_size": 0 00:25:07.045 }, 00:25:07.045 { 00:25:07.045 "name": "BaseBdev3", 00:25:07.045 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:07.045 "is_configured": false, 00:25:07.045 "data_offset": 0, 00:25:07.045 "data_size": 0 00:25:07.045 }, 00:25:07.045 { 00:25:07.045 "name": "BaseBdev4", 00:25:07.045 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:07.045 "is_configured": false, 00:25:07.045 "data_offset": 0, 00:25:07.045 "data_size": 0 00:25:07.045 } 00:25:07.045 ] 00:25:07.045 }' 00:25:07.045 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:07.045 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:07.304 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:25:07.304 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:07.304 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:07.304 [2024-10-01 20:24:02.536230] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:25:07.304 [2024-10-01 20:24:02.536318] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:25:07.304 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:07.304 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:25:07.304 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:07.304 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:07.304 [2024-10-01 20:24:02.544221] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:25:07.304 [2024-10-01 20:24:02.544273] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:25:07.304 [2024-10-01 20:24:02.544289] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:25:07.304 [2024-10-01 20:24:02.544306] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:25:07.304 [2024-10-01 20:24:02.544317] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:25:07.304 [2024-10-01 20:24:02.544333] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:25:07.304 [2024-10-01 20:24:02.544343] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:25:07.304 [2024-10-01 20:24:02.544359] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:25:07.304 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:07.304 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:25:07.304 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:07.304 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:07.563 [2024-10-01 20:24:02.590420] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:25:07.563 BaseBdev1 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:07.563 [ 00:25:07.563 { 00:25:07.563 "name": "BaseBdev1", 00:25:07.563 "aliases": [ 00:25:07.563 "d0215a7e-7eba-4e90-bd85-266a971e6a9e" 00:25:07.563 ], 00:25:07.563 "product_name": "Malloc disk", 00:25:07.563 "block_size": 512, 00:25:07.563 "num_blocks": 65536, 00:25:07.563 "uuid": "d0215a7e-7eba-4e90-bd85-266a971e6a9e", 00:25:07.563 "assigned_rate_limits": { 00:25:07.563 "rw_ios_per_sec": 0, 00:25:07.563 "rw_mbytes_per_sec": 0, 00:25:07.563 "r_mbytes_per_sec": 0, 00:25:07.563 "w_mbytes_per_sec": 0 00:25:07.563 }, 00:25:07.563 "claimed": true, 00:25:07.563 "claim_type": "exclusive_write", 00:25:07.563 "zoned": false, 00:25:07.563 "supported_io_types": { 00:25:07.563 "read": true, 00:25:07.563 "write": true, 00:25:07.563 "unmap": true, 00:25:07.563 "flush": true, 00:25:07.563 "reset": true, 00:25:07.563 "nvme_admin": false, 00:25:07.563 "nvme_io": false, 00:25:07.563 "nvme_io_md": false, 00:25:07.563 "write_zeroes": true, 00:25:07.563 "zcopy": true, 00:25:07.563 "get_zone_info": false, 00:25:07.563 "zone_management": false, 00:25:07.563 "zone_append": false, 00:25:07.563 "compare": false, 00:25:07.563 "compare_and_write": false, 00:25:07.563 "abort": true, 00:25:07.563 "seek_hole": false, 00:25:07.563 "seek_data": false, 00:25:07.563 "copy": true, 00:25:07.563 "nvme_iov_md": false 00:25:07.563 }, 00:25:07.563 "memory_domains": [ 00:25:07.563 { 00:25:07.563 "dma_device_id": "system", 00:25:07.563 "dma_device_type": 1 00:25:07.563 }, 00:25:07.563 { 00:25:07.563 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:07.563 "dma_device_type": 2 00:25:07.563 } 00:25:07.563 ], 00:25:07.563 "driver_specific": {} 00:25:07.563 } 00:25:07.563 ] 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:07.563 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:07.563 "name": "Existed_Raid", 00:25:07.563 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:07.563 "strip_size_kb": 0, 00:25:07.563 "state": "configuring", 00:25:07.563 "raid_level": "raid1", 00:25:07.563 "superblock": false, 00:25:07.563 "num_base_bdevs": 4, 00:25:07.563 "num_base_bdevs_discovered": 1, 00:25:07.563 "num_base_bdevs_operational": 4, 00:25:07.563 "base_bdevs_list": [ 00:25:07.563 { 00:25:07.563 "name": "BaseBdev1", 00:25:07.563 "uuid": "d0215a7e-7eba-4e90-bd85-266a971e6a9e", 00:25:07.563 "is_configured": true, 00:25:07.563 "data_offset": 0, 00:25:07.563 "data_size": 65536 00:25:07.563 }, 00:25:07.563 { 00:25:07.563 "name": "BaseBdev2", 00:25:07.563 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:07.563 "is_configured": false, 00:25:07.563 "data_offset": 0, 00:25:07.564 "data_size": 0 00:25:07.564 }, 00:25:07.564 { 00:25:07.564 "name": "BaseBdev3", 00:25:07.564 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:07.564 "is_configured": false, 00:25:07.564 "data_offset": 0, 00:25:07.564 "data_size": 0 00:25:07.564 }, 00:25:07.564 { 00:25:07.564 "name": "BaseBdev4", 00:25:07.564 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:07.564 "is_configured": false, 00:25:07.564 "data_offset": 0, 00:25:07.564 "data_size": 0 00:25:07.564 } 00:25:07.564 ] 00:25:07.564 }' 00:25:07.564 20:24:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:07.564 20:24:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:08.132 [2024-10-01 20:24:03.154627] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:25:08.132 [2024-10-01 20:24:03.154704] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:08.132 [2024-10-01 20:24:03.166671] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:25:08.132 [2024-10-01 20:24:03.169253] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:25:08.132 [2024-10-01 20:24:03.169314] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:25:08.132 [2024-10-01 20:24:03.169332] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:25:08.132 [2024-10-01 20:24:03.169352] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:25:08.132 [2024-10-01 20:24:03.169363] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:25:08.132 [2024-10-01 20:24:03.169378] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:08.132 "name": "Existed_Raid", 00:25:08.132 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:08.132 "strip_size_kb": 0, 00:25:08.132 "state": "configuring", 00:25:08.132 "raid_level": "raid1", 00:25:08.132 "superblock": false, 00:25:08.132 "num_base_bdevs": 4, 00:25:08.132 "num_base_bdevs_discovered": 1, 00:25:08.132 "num_base_bdevs_operational": 4, 00:25:08.132 "base_bdevs_list": [ 00:25:08.132 { 00:25:08.132 "name": "BaseBdev1", 00:25:08.132 "uuid": "d0215a7e-7eba-4e90-bd85-266a971e6a9e", 00:25:08.132 "is_configured": true, 00:25:08.132 "data_offset": 0, 00:25:08.132 "data_size": 65536 00:25:08.132 }, 00:25:08.132 { 00:25:08.132 "name": "BaseBdev2", 00:25:08.132 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:08.132 "is_configured": false, 00:25:08.132 "data_offset": 0, 00:25:08.132 "data_size": 0 00:25:08.132 }, 00:25:08.132 { 00:25:08.132 "name": "BaseBdev3", 00:25:08.132 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:08.132 "is_configured": false, 00:25:08.132 "data_offset": 0, 00:25:08.132 "data_size": 0 00:25:08.132 }, 00:25:08.132 { 00:25:08.132 "name": "BaseBdev4", 00:25:08.132 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:08.132 "is_configured": false, 00:25:08.132 "data_offset": 0, 00:25:08.132 "data_size": 0 00:25:08.132 } 00:25:08.132 ] 00:25:08.132 }' 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:08.132 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:08.700 [2024-10-01 20:24:03.726233] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:25:08.700 BaseBdev2 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:08.700 [ 00:25:08.700 { 00:25:08.700 "name": "BaseBdev2", 00:25:08.700 "aliases": [ 00:25:08.700 "e0a1e686-e068-445c-b394-a00cff909773" 00:25:08.700 ], 00:25:08.700 "product_name": "Malloc disk", 00:25:08.700 "block_size": 512, 00:25:08.700 "num_blocks": 65536, 00:25:08.700 "uuid": "e0a1e686-e068-445c-b394-a00cff909773", 00:25:08.700 "assigned_rate_limits": { 00:25:08.700 "rw_ios_per_sec": 0, 00:25:08.700 "rw_mbytes_per_sec": 0, 00:25:08.700 "r_mbytes_per_sec": 0, 00:25:08.700 "w_mbytes_per_sec": 0 00:25:08.700 }, 00:25:08.700 "claimed": true, 00:25:08.700 "claim_type": "exclusive_write", 00:25:08.700 "zoned": false, 00:25:08.700 "supported_io_types": { 00:25:08.700 "read": true, 00:25:08.700 "write": true, 00:25:08.700 "unmap": true, 00:25:08.700 "flush": true, 00:25:08.700 "reset": true, 00:25:08.700 "nvme_admin": false, 00:25:08.700 "nvme_io": false, 00:25:08.700 "nvme_io_md": false, 00:25:08.700 "write_zeroes": true, 00:25:08.700 "zcopy": true, 00:25:08.700 "get_zone_info": false, 00:25:08.700 "zone_management": false, 00:25:08.700 "zone_append": false, 00:25:08.700 "compare": false, 00:25:08.700 "compare_and_write": false, 00:25:08.700 "abort": true, 00:25:08.700 "seek_hole": false, 00:25:08.700 "seek_data": false, 00:25:08.700 "copy": true, 00:25:08.700 "nvme_iov_md": false 00:25:08.700 }, 00:25:08.700 "memory_domains": [ 00:25:08.700 { 00:25:08.700 "dma_device_id": "system", 00:25:08.700 "dma_device_type": 1 00:25:08.700 }, 00:25:08.700 { 00:25:08.700 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:08.700 "dma_device_type": 2 00:25:08.700 } 00:25:08.700 ], 00:25:08.700 "driver_specific": {} 00:25:08.700 } 00:25:08.700 ] 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:08.700 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:08.700 "name": "Existed_Raid", 00:25:08.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:08.700 "strip_size_kb": 0, 00:25:08.700 "state": "configuring", 00:25:08.700 "raid_level": "raid1", 00:25:08.700 "superblock": false, 00:25:08.700 "num_base_bdevs": 4, 00:25:08.700 "num_base_bdevs_discovered": 2, 00:25:08.700 "num_base_bdevs_operational": 4, 00:25:08.700 "base_bdevs_list": [ 00:25:08.700 { 00:25:08.700 "name": "BaseBdev1", 00:25:08.700 "uuid": "d0215a7e-7eba-4e90-bd85-266a971e6a9e", 00:25:08.700 "is_configured": true, 00:25:08.700 "data_offset": 0, 00:25:08.700 "data_size": 65536 00:25:08.700 }, 00:25:08.700 { 00:25:08.700 "name": "BaseBdev2", 00:25:08.700 "uuid": "e0a1e686-e068-445c-b394-a00cff909773", 00:25:08.700 "is_configured": true, 00:25:08.700 "data_offset": 0, 00:25:08.700 "data_size": 65536 00:25:08.701 }, 00:25:08.701 { 00:25:08.701 "name": "BaseBdev3", 00:25:08.701 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:08.701 "is_configured": false, 00:25:08.701 "data_offset": 0, 00:25:08.701 "data_size": 0 00:25:08.701 }, 00:25:08.701 { 00:25:08.701 "name": "BaseBdev4", 00:25:08.701 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:08.701 "is_configured": false, 00:25:08.701 "data_offset": 0, 00:25:08.701 "data_size": 0 00:25:08.701 } 00:25:08.701 ] 00:25:08.701 }' 00:25:08.701 20:24:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:08.701 20:24:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:09.269 [2024-10-01 20:24:04.329912] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:25:09.269 BaseBdev3 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:09.269 [ 00:25:09.269 { 00:25:09.269 "name": "BaseBdev3", 00:25:09.269 "aliases": [ 00:25:09.269 "25696398-6d77-4450-a42b-5f6faca4ac5c" 00:25:09.269 ], 00:25:09.269 "product_name": "Malloc disk", 00:25:09.269 "block_size": 512, 00:25:09.269 "num_blocks": 65536, 00:25:09.269 "uuid": "25696398-6d77-4450-a42b-5f6faca4ac5c", 00:25:09.269 "assigned_rate_limits": { 00:25:09.269 "rw_ios_per_sec": 0, 00:25:09.269 "rw_mbytes_per_sec": 0, 00:25:09.269 "r_mbytes_per_sec": 0, 00:25:09.269 "w_mbytes_per_sec": 0 00:25:09.269 }, 00:25:09.269 "claimed": true, 00:25:09.269 "claim_type": "exclusive_write", 00:25:09.269 "zoned": false, 00:25:09.269 "supported_io_types": { 00:25:09.269 "read": true, 00:25:09.269 "write": true, 00:25:09.269 "unmap": true, 00:25:09.269 "flush": true, 00:25:09.269 "reset": true, 00:25:09.269 "nvme_admin": false, 00:25:09.269 "nvme_io": false, 00:25:09.269 "nvme_io_md": false, 00:25:09.269 "write_zeroes": true, 00:25:09.269 "zcopy": true, 00:25:09.269 "get_zone_info": false, 00:25:09.269 "zone_management": false, 00:25:09.269 "zone_append": false, 00:25:09.269 "compare": false, 00:25:09.269 "compare_and_write": false, 00:25:09.269 "abort": true, 00:25:09.269 "seek_hole": false, 00:25:09.269 "seek_data": false, 00:25:09.269 "copy": true, 00:25:09.269 "nvme_iov_md": false 00:25:09.269 }, 00:25:09.269 "memory_domains": [ 00:25:09.269 { 00:25:09.269 "dma_device_id": "system", 00:25:09.269 "dma_device_type": 1 00:25:09.269 }, 00:25:09.269 { 00:25:09.269 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:09.269 "dma_device_type": 2 00:25:09.269 } 00:25:09.269 ], 00:25:09.269 "driver_specific": {} 00:25:09.269 } 00:25:09.269 ] 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:09.269 "name": "Existed_Raid", 00:25:09.269 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:09.269 "strip_size_kb": 0, 00:25:09.269 "state": "configuring", 00:25:09.269 "raid_level": "raid1", 00:25:09.269 "superblock": false, 00:25:09.269 "num_base_bdevs": 4, 00:25:09.269 "num_base_bdevs_discovered": 3, 00:25:09.269 "num_base_bdevs_operational": 4, 00:25:09.269 "base_bdevs_list": [ 00:25:09.269 { 00:25:09.269 "name": "BaseBdev1", 00:25:09.269 "uuid": "d0215a7e-7eba-4e90-bd85-266a971e6a9e", 00:25:09.269 "is_configured": true, 00:25:09.269 "data_offset": 0, 00:25:09.269 "data_size": 65536 00:25:09.269 }, 00:25:09.269 { 00:25:09.269 "name": "BaseBdev2", 00:25:09.269 "uuid": "e0a1e686-e068-445c-b394-a00cff909773", 00:25:09.269 "is_configured": true, 00:25:09.269 "data_offset": 0, 00:25:09.269 "data_size": 65536 00:25:09.269 }, 00:25:09.269 { 00:25:09.269 "name": "BaseBdev3", 00:25:09.269 "uuid": "25696398-6d77-4450-a42b-5f6faca4ac5c", 00:25:09.269 "is_configured": true, 00:25:09.269 "data_offset": 0, 00:25:09.269 "data_size": 65536 00:25:09.269 }, 00:25:09.269 { 00:25:09.269 "name": "BaseBdev4", 00:25:09.269 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:09.269 "is_configured": false, 00:25:09.269 "data_offset": 0, 00:25:09.269 "data_size": 0 00:25:09.269 } 00:25:09.269 ] 00:25:09.269 }' 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:09.269 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:09.837 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:25:09.837 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:09.837 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:09.837 [2024-10-01 20:24:04.913115] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:25:09.837 [2024-10-01 20:24:04.913196] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:25:09.837 [2024-10-01 20:24:04.913216] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:25:09.837 [2024-10-01 20:24:04.913600] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:25:09.837 [2024-10-01 20:24:04.913868] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:25:09.837 [2024-10-01 20:24:04.913902] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:25:09.837 [2024-10-01 20:24:04.914255] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:09.837 BaseBdev4 00:25:09.837 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:09.838 [ 00:25:09.838 { 00:25:09.838 "name": "BaseBdev4", 00:25:09.838 "aliases": [ 00:25:09.838 "b16e6713-3122-48fe-a9dd-f50727cef77e" 00:25:09.838 ], 00:25:09.838 "product_name": "Malloc disk", 00:25:09.838 "block_size": 512, 00:25:09.838 "num_blocks": 65536, 00:25:09.838 "uuid": "b16e6713-3122-48fe-a9dd-f50727cef77e", 00:25:09.838 "assigned_rate_limits": { 00:25:09.838 "rw_ios_per_sec": 0, 00:25:09.838 "rw_mbytes_per_sec": 0, 00:25:09.838 "r_mbytes_per_sec": 0, 00:25:09.838 "w_mbytes_per_sec": 0 00:25:09.838 }, 00:25:09.838 "claimed": true, 00:25:09.838 "claim_type": "exclusive_write", 00:25:09.838 "zoned": false, 00:25:09.838 "supported_io_types": { 00:25:09.838 "read": true, 00:25:09.838 "write": true, 00:25:09.838 "unmap": true, 00:25:09.838 "flush": true, 00:25:09.838 "reset": true, 00:25:09.838 "nvme_admin": false, 00:25:09.838 "nvme_io": false, 00:25:09.838 "nvme_io_md": false, 00:25:09.838 "write_zeroes": true, 00:25:09.838 "zcopy": true, 00:25:09.838 "get_zone_info": false, 00:25:09.838 "zone_management": false, 00:25:09.838 "zone_append": false, 00:25:09.838 "compare": false, 00:25:09.838 "compare_and_write": false, 00:25:09.838 "abort": true, 00:25:09.838 "seek_hole": false, 00:25:09.838 "seek_data": false, 00:25:09.838 "copy": true, 00:25:09.838 "nvme_iov_md": false 00:25:09.838 }, 00:25:09.838 "memory_domains": [ 00:25:09.838 { 00:25:09.838 "dma_device_id": "system", 00:25:09.838 "dma_device_type": 1 00:25:09.838 }, 00:25:09.838 { 00:25:09.838 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:09.838 "dma_device_type": 2 00:25:09.838 } 00:25:09.838 ], 00:25:09.838 "driver_specific": {} 00:25:09.838 } 00:25:09.838 ] 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:09.838 20:24:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:09.838 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:09.838 "name": "Existed_Raid", 00:25:09.838 "uuid": "8244e479-8f5d-4990-a60d-9c9aecb90b32", 00:25:09.838 "strip_size_kb": 0, 00:25:09.838 "state": "online", 00:25:09.838 "raid_level": "raid1", 00:25:09.838 "superblock": false, 00:25:09.838 "num_base_bdevs": 4, 00:25:09.838 "num_base_bdevs_discovered": 4, 00:25:09.838 "num_base_bdevs_operational": 4, 00:25:09.838 "base_bdevs_list": [ 00:25:09.838 { 00:25:09.838 "name": "BaseBdev1", 00:25:09.838 "uuid": "d0215a7e-7eba-4e90-bd85-266a971e6a9e", 00:25:09.838 "is_configured": true, 00:25:09.838 "data_offset": 0, 00:25:09.838 "data_size": 65536 00:25:09.838 }, 00:25:09.838 { 00:25:09.838 "name": "BaseBdev2", 00:25:09.838 "uuid": "e0a1e686-e068-445c-b394-a00cff909773", 00:25:09.838 "is_configured": true, 00:25:09.838 "data_offset": 0, 00:25:09.838 "data_size": 65536 00:25:09.838 }, 00:25:09.838 { 00:25:09.838 "name": "BaseBdev3", 00:25:09.838 "uuid": "25696398-6d77-4450-a42b-5f6faca4ac5c", 00:25:09.838 "is_configured": true, 00:25:09.838 "data_offset": 0, 00:25:09.838 "data_size": 65536 00:25:09.838 }, 00:25:09.838 { 00:25:09.838 "name": "BaseBdev4", 00:25:09.838 "uuid": "b16e6713-3122-48fe-a9dd-f50727cef77e", 00:25:09.838 "is_configured": true, 00:25:09.838 "data_offset": 0, 00:25:09.838 "data_size": 65536 00:25:09.838 } 00:25:09.838 ] 00:25:09.838 }' 00:25:09.838 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:09.838 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:10.406 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:25:10.406 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:25:10.406 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:25:10.406 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:25:10.406 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:25:10.406 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:25:10.406 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:25:10.406 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:25:10.406 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:10.406 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:10.406 [2024-10-01 20:24:05.473780] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:25:10.406 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:10.406 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:10.406 "name": "Existed_Raid", 00:25:10.406 "aliases": [ 00:25:10.406 "8244e479-8f5d-4990-a60d-9c9aecb90b32" 00:25:10.406 ], 00:25:10.406 "product_name": "Raid Volume", 00:25:10.406 "block_size": 512, 00:25:10.406 "num_blocks": 65536, 00:25:10.406 "uuid": "8244e479-8f5d-4990-a60d-9c9aecb90b32", 00:25:10.406 "assigned_rate_limits": { 00:25:10.406 "rw_ios_per_sec": 0, 00:25:10.406 "rw_mbytes_per_sec": 0, 00:25:10.406 "r_mbytes_per_sec": 0, 00:25:10.406 "w_mbytes_per_sec": 0 00:25:10.406 }, 00:25:10.406 "claimed": false, 00:25:10.406 "zoned": false, 00:25:10.406 "supported_io_types": { 00:25:10.406 "read": true, 00:25:10.406 "write": true, 00:25:10.406 "unmap": false, 00:25:10.406 "flush": false, 00:25:10.406 "reset": true, 00:25:10.406 "nvme_admin": false, 00:25:10.406 "nvme_io": false, 00:25:10.406 "nvme_io_md": false, 00:25:10.406 "write_zeroes": true, 00:25:10.406 "zcopy": false, 00:25:10.406 "get_zone_info": false, 00:25:10.406 "zone_management": false, 00:25:10.406 "zone_append": false, 00:25:10.406 "compare": false, 00:25:10.406 "compare_and_write": false, 00:25:10.406 "abort": false, 00:25:10.406 "seek_hole": false, 00:25:10.406 "seek_data": false, 00:25:10.406 "copy": false, 00:25:10.406 "nvme_iov_md": false 00:25:10.406 }, 00:25:10.406 "memory_domains": [ 00:25:10.406 { 00:25:10.406 "dma_device_id": "system", 00:25:10.406 "dma_device_type": 1 00:25:10.406 }, 00:25:10.406 { 00:25:10.406 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:10.406 "dma_device_type": 2 00:25:10.406 }, 00:25:10.406 { 00:25:10.406 "dma_device_id": "system", 00:25:10.406 "dma_device_type": 1 00:25:10.406 }, 00:25:10.406 { 00:25:10.406 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:10.406 "dma_device_type": 2 00:25:10.406 }, 00:25:10.406 { 00:25:10.406 "dma_device_id": "system", 00:25:10.406 "dma_device_type": 1 00:25:10.406 }, 00:25:10.406 { 00:25:10.406 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:10.406 "dma_device_type": 2 00:25:10.406 }, 00:25:10.406 { 00:25:10.406 "dma_device_id": "system", 00:25:10.406 "dma_device_type": 1 00:25:10.406 }, 00:25:10.406 { 00:25:10.406 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:10.406 "dma_device_type": 2 00:25:10.406 } 00:25:10.406 ], 00:25:10.406 "driver_specific": { 00:25:10.406 "raid": { 00:25:10.406 "uuid": "8244e479-8f5d-4990-a60d-9c9aecb90b32", 00:25:10.406 "strip_size_kb": 0, 00:25:10.406 "state": "online", 00:25:10.406 "raid_level": "raid1", 00:25:10.406 "superblock": false, 00:25:10.406 "num_base_bdevs": 4, 00:25:10.406 "num_base_bdevs_discovered": 4, 00:25:10.406 "num_base_bdevs_operational": 4, 00:25:10.406 "base_bdevs_list": [ 00:25:10.406 { 00:25:10.406 "name": "BaseBdev1", 00:25:10.406 "uuid": "d0215a7e-7eba-4e90-bd85-266a971e6a9e", 00:25:10.406 "is_configured": true, 00:25:10.406 "data_offset": 0, 00:25:10.406 "data_size": 65536 00:25:10.406 }, 00:25:10.406 { 00:25:10.406 "name": "BaseBdev2", 00:25:10.406 "uuid": "e0a1e686-e068-445c-b394-a00cff909773", 00:25:10.406 "is_configured": true, 00:25:10.406 "data_offset": 0, 00:25:10.406 "data_size": 65536 00:25:10.406 }, 00:25:10.406 { 00:25:10.406 "name": "BaseBdev3", 00:25:10.406 "uuid": "25696398-6d77-4450-a42b-5f6faca4ac5c", 00:25:10.406 "is_configured": true, 00:25:10.406 "data_offset": 0, 00:25:10.406 "data_size": 65536 00:25:10.406 }, 00:25:10.406 { 00:25:10.406 "name": "BaseBdev4", 00:25:10.406 "uuid": "b16e6713-3122-48fe-a9dd-f50727cef77e", 00:25:10.406 "is_configured": true, 00:25:10.406 "data_offset": 0, 00:25:10.406 "data_size": 65536 00:25:10.406 } 00:25:10.406 ] 00:25:10.406 } 00:25:10.406 } 00:25:10.406 }' 00:25:10.406 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:25:10.406 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:25:10.406 BaseBdev2 00:25:10.406 BaseBdev3 00:25:10.406 BaseBdev4' 00:25:10.406 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:10.407 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:25:10.407 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:10.407 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:25:10.407 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:10.407 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:10.407 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:10.407 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:10.665 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:10.666 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:10.666 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:10.666 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:25:10.666 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:10.666 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:10.666 [2024-10-01 20:24:05.829514] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:25:10.666 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:10.666 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:25:10.666 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:25:10.666 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:25:10.666 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:10.924 "name": "Existed_Raid", 00:25:10.924 "uuid": "8244e479-8f5d-4990-a60d-9c9aecb90b32", 00:25:10.924 "strip_size_kb": 0, 00:25:10.924 "state": "online", 00:25:10.924 "raid_level": "raid1", 00:25:10.924 "superblock": false, 00:25:10.924 "num_base_bdevs": 4, 00:25:10.924 "num_base_bdevs_discovered": 3, 00:25:10.924 "num_base_bdevs_operational": 3, 00:25:10.924 "base_bdevs_list": [ 00:25:10.924 { 00:25:10.924 "name": null, 00:25:10.924 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:10.924 "is_configured": false, 00:25:10.924 "data_offset": 0, 00:25:10.924 "data_size": 65536 00:25:10.924 }, 00:25:10.924 { 00:25:10.924 "name": "BaseBdev2", 00:25:10.924 "uuid": "e0a1e686-e068-445c-b394-a00cff909773", 00:25:10.924 "is_configured": true, 00:25:10.924 "data_offset": 0, 00:25:10.924 "data_size": 65536 00:25:10.924 }, 00:25:10.924 { 00:25:10.924 "name": "BaseBdev3", 00:25:10.924 "uuid": "25696398-6d77-4450-a42b-5f6faca4ac5c", 00:25:10.924 "is_configured": true, 00:25:10.924 "data_offset": 0, 00:25:10.924 "data_size": 65536 00:25:10.924 }, 00:25:10.924 { 00:25:10.924 "name": "BaseBdev4", 00:25:10.924 "uuid": "b16e6713-3122-48fe-a9dd-f50727cef77e", 00:25:10.924 "is_configured": true, 00:25:10.924 "data_offset": 0, 00:25:10.924 "data_size": 65536 00:25:10.924 } 00:25:10.924 ] 00:25:10.924 }' 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:10.924 20:24:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.182 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:25:11.182 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.441 [2024-10-01 20:24:06.492287] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.441 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.441 [2024-10-01 20:24:06.639969] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.701 [2024-10-01 20:24:06.785455] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:25:11.701 [2024-10-01 20:24:06.785584] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:25:11.701 [2024-10-01 20:24:06.874434] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:25:11.701 [2024-10-01 20:24:06.874526] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:25:11.701 [2024-10-01 20:24:06.874549] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.701 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.960 BaseBdev2 00:25:11.960 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.960 20:24:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:25:11.960 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:25:11.960 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:11.960 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:25:11.960 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:11.960 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:11.960 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:11.960 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.960 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.960 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.960 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:25:11.960 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.960 20:24:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.960 [ 00:25:11.960 { 00:25:11.960 "name": "BaseBdev2", 00:25:11.960 "aliases": [ 00:25:11.960 "c092e1b1-4405-414f-8dbe-67267435512c" 00:25:11.960 ], 00:25:11.960 "product_name": "Malloc disk", 00:25:11.960 "block_size": 512, 00:25:11.960 "num_blocks": 65536, 00:25:11.960 "uuid": "c092e1b1-4405-414f-8dbe-67267435512c", 00:25:11.960 "assigned_rate_limits": { 00:25:11.960 "rw_ios_per_sec": 0, 00:25:11.960 "rw_mbytes_per_sec": 0, 00:25:11.960 "r_mbytes_per_sec": 0, 00:25:11.960 "w_mbytes_per_sec": 0 00:25:11.960 }, 00:25:11.960 "claimed": false, 00:25:11.960 "zoned": false, 00:25:11.960 "supported_io_types": { 00:25:11.960 "read": true, 00:25:11.960 "write": true, 00:25:11.960 "unmap": true, 00:25:11.960 "flush": true, 00:25:11.960 "reset": true, 00:25:11.960 "nvme_admin": false, 00:25:11.960 "nvme_io": false, 00:25:11.960 "nvme_io_md": false, 00:25:11.960 "write_zeroes": true, 00:25:11.960 "zcopy": true, 00:25:11.960 "get_zone_info": false, 00:25:11.960 "zone_management": false, 00:25:11.960 "zone_append": false, 00:25:11.960 "compare": false, 00:25:11.960 "compare_and_write": false, 00:25:11.960 "abort": true, 00:25:11.960 "seek_hole": false, 00:25:11.960 "seek_data": false, 00:25:11.960 "copy": true, 00:25:11.960 "nvme_iov_md": false 00:25:11.960 }, 00:25:11.960 "memory_domains": [ 00:25:11.960 { 00:25:11.960 "dma_device_id": "system", 00:25:11.960 "dma_device_type": 1 00:25:11.960 }, 00:25:11.960 { 00:25:11.960 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:11.960 "dma_device_type": 2 00:25:11.960 } 00:25:11.960 ], 00:25:11.960 "driver_specific": {} 00:25:11.960 } 00:25:11.960 ] 00:25:11.960 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.961 BaseBdev3 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.961 [ 00:25:11.961 { 00:25:11.961 "name": "BaseBdev3", 00:25:11.961 "aliases": [ 00:25:11.961 "dce527a3-a7b0-4407-ae6b-9e331428733a" 00:25:11.961 ], 00:25:11.961 "product_name": "Malloc disk", 00:25:11.961 "block_size": 512, 00:25:11.961 "num_blocks": 65536, 00:25:11.961 "uuid": "dce527a3-a7b0-4407-ae6b-9e331428733a", 00:25:11.961 "assigned_rate_limits": { 00:25:11.961 "rw_ios_per_sec": 0, 00:25:11.961 "rw_mbytes_per_sec": 0, 00:25:11.961 "r_mbytes_per_sec": 0, 00:25:11.961 "w_mbytes_per_sec": 0 00:25:11.961 }, 00:25:11.961 "claimed": false, 00:25:11.961 "zoned": false, 00:25:11.961 "supported_io_types": { 00:25:11.961 "read": true, 00:25:11.961 "write": true, 00:25:11.961 "unmap": true, 00:25:11.961 "flush": true, 00:25:11.961 "reset": true, 00:25:11.961 "nvme_admin": false, 00:25:11.961 "nvme_io": false, 00:25:11.961 "nvme_io_md": false, 00:25:11.961 "write_zeroes": true, 00:25:11.961 "zcopy": true, 00:25:11.961 "get_zone_info": false, 00:25:11.961 "zone_management": false, 00:25:11.961 "zone_append": false, 00:25:11.961 "compare": false, 00:25:11.961 "compare_and_write": false, 00:25:11.961 "abort": true, 00:25:11.961 "seek_hole": false, 00:25:11.961 "seek_data": false, 00:25:11.961 "copy": true, 00:25:11.961 "nvme_iov_md": false 00:25:11.961 }, 00:25:11.961 "memory_domains": [ 00:25:11.961 { 00:25:11.961 "dma_device_id": "system", 00:25:11.961 "dma_device_type": 1 00:25:11.961 }, 00:25:11.961 { 00:25:11.961 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:11.961 "dma_device_type": 2 00:25:11.961 } 00:25:11.961 ], 00:25:11.961 "driver_specific": {} 00:25:11.961 } 00:25:11.961 ] 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.961 BaseBdev4 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.961 [ 00:25:11.961 { 00:25:11.961 "name": "BaseBdev4", 00:25:11.961 "aliases": [ 00:25:11.961 "6f1779e0-4d39-4734-83e1-4c8ac9520198" 00:25:11.961 ], 00:25:11.961 "product_name": "Malloc disk", 00:25:11.961 "block_size": 512, 00:25:11.961 "num_blocks": 65536, 00:25:11.961 "uuid": "6f1779e0-4d39-4734-83e1-4c8ac9520198", 00:25:11.961 "assigned_rate_limits": { 00:25:11.961 "rw_ios_per_sec": 0, 00:25:11.961 "rw_mbytes_per_sec": 0, 00:25:11.961 "r_mbytes_per_sec": 0, 00:25:11.961 "w_mbytes_per_sec": 0 00:25:11.961 }, 00:25:11.961 "claimed": false, 00:25:11.961 "zoned": false, 00:25:11.961 "supported_io_types": { 00:25:11.961 "read": true, 00:25:11.961 "write": true, 00:25:11.961 "unmap": true, 00:25:11.961 "flush": true, 00:25:11.961 "reset": true, 00:25:11.961 "nvme_admin": false, 00:25:11.961 "nvme_io": false, 00:25:11.961 "nvme_io_md": false, 00:25:11.961 "write_zeroes": true, 00:25:11.961 "zcopy": true, 00:25:11.961 "get_zone_info": false, 00:25:11.961 "zone_management": false, 00:25:11.961 "zone_append": false, 00:25:11.961 "compare": false, 00:25:11.961 "compare_and_write": false, 00:25:11.961 "abort": true, 00:25:11.961 "seek_hole": false, 00:25:11.961 "seek_data": false, 00:25:11.961 "copy": true, 00:25:11.961 "nvme_iov_md": false 00:25:11.961 }, 00:25:11.961 "memory_domains": [ 00:25:11.961 { 00:25:11.961 "dma_device_id": "system", 00:25:11.961 "dma_device_type": 1 00:25:11.961 }, 00:25:11.961 { 00:25:11.961 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:11.961 "dma_device_type": 2 00:25:11.961 } 00:25:11.961 ], 00:25:11.961 "driver_specific": {} 00:25:11.961 } 00:25:11.961 ] 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.961 [2024-10-01 20:24:07.157070] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:25:11.961 [2024-10-01 20:24:07.157135] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:25:11.961 [2024-10-01 20:24:07.157164] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:25:11.961 [2024-10-01 20:24:07.159642] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:25:11.961 [2024-10-01 20:24:07.159732] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:11.961 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:11.961 "name": "Existed_Raid", 00:25:11.961 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:11.961 "strip_size_kb": 0, 00:25:11.961 "state": "configuring", 00:25:11.961 "raid_level": "raid1", 00:25:11.962 "superblock": false, 00:25:11.962 "num_base_bdevs": 4, 00:25:11.962 "num_base_bdevs_discovered": 3, 00:25:11.962 "num_base_bdevs_operational": 4, 00:25:11.962 "base_bdevs_list": [ 00:25:11.962 { 00:25:11.962 "name": "BaseBdev1", 00:25:11.962 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:11.962 "is_configured": false, 00:25:11.962 "data_offset": 0, 00:25:11.962 "data_size": 0 00:25:11.962 }, 00:25:11.962 { 00:25:11.962 "name": "BaseBdev2", 00:25:11.962 "uuid": "c092e1b1-4405-414f-8dbe-67267435512c", 00:25:11.962 "is_configured": true, 00:25:11.962 "data_offset": 0, 00:25:11.962 "data_size": 65536 00:25:11.962 }, 00:25:11.962 { 00:25:11.962 "name": "BaseBdev3", 00:25:11.962 "uuid": "dce527a3-a7b0-4407-ae6b-9e331428733a", 00:25:11.962 "is_configured": true, 00:25:11.962 "data_offset": 0, 00:25:11.962 "data_size": 65536 00:25:11.962 }, 00:25:11.962 { 00:25:11.962 "name": "BaseBdev4", 00:25:11.962 "uuid": "6f1779e0-4d39-4734-83e1-4c8ac9520198", 00:25:11.962 "is_configured": true, 00:25:11.962 "data_offset": 0, 00:25:11.962 "data_size": 65536 00:25:11.962 } 00:25:11.962 ] 00:25:11.962 }' 00:25:11.962 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:11.962 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:12.529 [2024-10-01 20:24:07.661262] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:12.529 "name": "Existed_Raid", 00:25:12.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:12.529 "strip_size_kb": 0, 00:25:12.529 "state": "configuring", 00:25:12.529 "raid_level": "raid1", 00:25:12.529 "superblock": false, 00:25:12.529 "num_base_bdevs": 4, 00:25:12.529 "num_base_bdevs_discovered": 2, 00:25:12.529 "num_base_bdevs_operational": 4, 00:25:12.529 "base_bdevs_list": [ 00:25:12.529 { 00:25:12.529 "name": "BaseBdev1", 00:25:12.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:12.529 "is_configured": false, 00:25:12.529 "data_offset": 0, 00:25:12.529 "data_size": 0 00:25:12.529 }, 00:25:12.529 { 00:25:12.529 "name": null, 00:25:12.529 "uuid": "c092e1b1-4405-414f-8dbe-67267435512c", 00:25:12.529 "is_configured": false, 00:25:12.529 "data_offset": 0, 00:25:12.529 "data_size": 65536 00:25:12.529 }, 00:25:12.529 { 00:25:12.529 "name": "BaseBdev3", 00:25:12.529 "uuid": "dce527a3-a7b0-4407-ae6b-9e331428733a", 00:25:12.529 "is_configured": true, 00:25:12.529 "data_offset": 0, 00:25:12.529 "data_size": 65536 00:25:12.529 }, 00:25:12.529 { 00:25:12.529 "name": "BaseBdev4", 00:25:12.529 "uuid": "6f1779e0-4d39-4734-83e1-4c8ac9520198", 00:25:12.529 "is_configured": true, 00:25:12.529 "data_offset": 0, 00:25:12.529 "data_size": 65536 00:25:12.529 } 00:25:12.529 ] 00:25:12.529 }' 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:12.529 20:24:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:13.156 [2024-10-01 20:24:08.279954] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:25:13.156 BaseBdev1 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:13.156 [ 00:25:13.156 { 00:25:13.156 "name": "BaseBdev1", 00:25:13.156 "aliases": [ 00:25:13.156 "daf00519-385f-4951-b4fd-fcf5cf3d8b74" 00:25:13.156 ], 00:25:13.156 "product_name": "Malloc disk", 00:25:13.156 "block_size": 512, 00:25:13.156 "num_blocks": 65536, 00:25:13.156 "uuid": "daf00519-385f-4951-b4fd-fcf5cf3d8b74", 00:25:13.156 "assigned_rate_limits": { 00:25:13.156 "rw_ios_per_sec": 0, 00:25:13.156 "rw_mbytes_per_sec": 0, 00:25:13.156 "r_mbytes_per_sec": 0, 00:25:13.156 "w_mbytes_per_sec": 0 00:25:13.156 }, 00:25:13.156 "claimed": true, 00:25:13.156 "claim_type": "exclusive_write", 00:25:13.156 "zoned": false, 00:25:13.156 "supported_io_types": { 00:25:13.156 "read": true, 00:25:13.156 "write": true, 00:25:13.156 "unmap": true, 00:25:13.156 "flush": true, 00:25:13.156 "reset": true, 00:25:13.156 "nvme_admin": false, 00:25:13.156 "nvme_io": false, 00:25:13.156 "nvme_io_md": false, 00:25:13.156 "write_zeroes": true, 00:25:13.156 "zcopy": true, 00:25:13.156 "get_zone_info": false, 00:25:13.156 "zone_management": false, 00:25:13.156 "zone_append": false, 00:25:13.156 "compare": false, 00:25:13.156 "compare_and_write": false, 00:25:13.156 "abort": true, 00:25:13.156 "seek_hole": false, 00:25:13.156 "seek_data": false, 00:25:13.156 "copy": true, 00:25:13.156 "nvme_iov_md": false 00:25:13.156 }, 00:25:13.156 "memory_domains": [ 00:25:13.156 { 00:25:13.156 "dma_device_id": "system", 00:25:13.156 "dma_device_type": 1 00:25:13.156 }, 00:25:13.156 { 00:25:13.156 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:13.156 "dma_device_type": 2 00:25:13.156 } 00:25:13.156 ], 00:25:13.156 "driver_specific": {} 00:25:13.156 } 00:25:13.156 ] 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:13.156 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:13.156 "name": "Existed_Raid", 00:25:13.156 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:13.156 "strip_size_kb": 0, 00:25:13.156 "state": "configuring", 00:25:13.156 "raid_level": "raid1", 00:25:13.156 "superblock": false, 00:25:13.156 "num_base_bdevs": 4, 00:25:13.156 "num_base_bdevs_discovered": 3, 00:25:13.156 "num_base_bdevs_operational": 4, 00:25:13.156 "base_bdevs_list": [ 00:25:13.156 { 00:25:13.156 "name": "BaseBdev1", 00:25:13.156 "uuid": "daf00519-385f-4951-b4fd-fcf5cf3d8b74", 00:25:13.156 "is_configured": true, 00:25:13.156 "data_offset": 0, 00:25:13.156 "data_size": 65536 00:25:13.156 }, 00:25:13.156 { 00:25:13.156 "name": null, 00:25:13.156 "uuid": "c092e1b1-4405-414f-8dbe-67267435512c", 00:25:13.156 "is_configured": false, 00:25:13.156 "data_offset": 0, 00:25:13.156 "data_size": 65536 00:25:13.156 }, 00:25:13.156 { 00:25:13.156 "name": "BaseBdev3", 00:25:13.157 "uuid": "dce527a3-a7b0-4407-ae6b-9e331428733a", 00:25:13.157 "is_configured": true, 00:25:13.157 "data_offset": 0, 00:25:13.157 "data_size": 65536 00:25:13.157 }, 00:25:13.157 { 00:25:13.157 "name": "BaseBdev4", 00:25:13.157 "uuid": "6f1779e0-4d39-4734-83e1-4c8ac9520198", 00:25:13.157 "is_configured": true, 00:25:13.157 "data_offset": 0, 00:25:13.157 "data_size": 65536 00:25:13.157 } 00:25:13.157 ] 00:25:13.157 }' 00:25:13.157 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:13.157 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:13.740 [2024-10-01 20:24:08.864201] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:13.740 "name": "Existed_Raid", 00:25:13.740 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:13.740 "strip_size_kb": 0, 00:25:13.740 "state": "configuring", 00:25:13.740 "raid_level": "raid1", 00:25:13.740 "superblock": false, 00:25:13.740 "num_base_bdevs": 4, 00:25:13.740 "num_base_bdevs_discovered": 2, 00:25:13.740 "num_base_bdevs_operational": 4, 00:25:13.740 "base_bdevs_list": [ 00:25:13.740 { 00:25:13.740 "name": "BaseBdev1", 00:25:13.740 "uuid": "daf00519-385f-4951-b4fd-fcf5cf3d8b74", 00:25:13.740 "is_configured": true, 00:25:13.740 "data_offset": 0, 00:25:13.740 "data_size": 65536 00:25:13.740 }, 00:25:13.740 { 00:25:13.740 "name": null, 00:25:13.740 "uuid": "c092e1b1-4405-414f-8dbe-67267435512c", 00:25:13.740 "is_configured": false, 00:25:13.740 "data_offset": 0, 00:25:13.740 "data_size": 65536 00:25:13.740 }, 00:25:13.740 { 00:25:13.740 "name": null, 00:25:13.740 "uuid": "dce527a3-a7b0-4407-ae6b-9e331428733a", 00:25:13.740 "is_configured": false, 00:25:13.740 "data_offset": 0, 00:25:13.740 "data_size": 65536 00:25:13.740 }, 00:25:13.740 { 00:25:13.740 "name": "BaseBdev4", 00:25:13.740 "uuid": "6f1779e0-4d39-4734-83e1-4c8ac9520198", 00:25:13.740 "is_configured": true, 00:25:13.740 "data_offset": 0, 00:25:13.740 "data_size": 65536 00:25:13.740 } 00:25:13.740 ] 00:25:13.740 }' 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:13.740 20:24:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:14.309 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:14.309 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:14.310 [2024-10-01 20:24:09.432480] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:14.310 "name": "Existed_Raid", 00:25:14.310 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:14.310 "strip_size_kb": 0, 00:25:14.310 "state": "configuring", 00:25:14.310 "raid_level": "raid1", 00:25:14.310 "superblock": false, 00:25:14.310 "num_base_bdevs": 4, 00:25:14.310 "num_base_bdevs_discovered": 3, 00:25:14.310 "num_base_bdevs_operational": 4, 00:25:14.310 "base_bdevs_list": [ 00:25:14.310 { 00:25:14.310 "name": "BaseBdev1", 00:25:14.310 "uuid": "daf00519-385f-4951-b4fd-fcf5cf3d8b74", 00:25:14.310 "is_configured": true, 00:25:14.310 "data_offset": 0, 00:25:14.310 "data_size": 65536 00:25:14.310 }, 00:25:14.310 { 00:25:14.310 "name": null, 00:25:14.310 "uuid": "c092e1b1-4405-414f-8dbe-67267435512c", 00:25:14.310 "is_configured": false, 00:25:14.310 "data_offset": 0, 00:25:14.310 "data_size": 65536 00:25:14.310 }, 00:25:14.310 { 00:25:14.310 "name": "BaseBdev3", 00:25:14.310 "uuid": "dce527a3-a7b0-4407-ae6b-9e331428733a", 00:25:14.310 "is_configured": true, 00:25:14.310 "data_offset": 0, 00:25:14.310 "data_size": 65536 00:25:14.310 }, 00:25:14.310 { 00:25:14.310 "name": "BaseBdev4", 00:25:14.310 "uuid": "6f1779e0-4d39-4734-83e1-4c8ac9520198", 00:25:14.310 "is_configured": true, 00:25:14.310 "data_offset": 0, 00:25:14.310 "data_size": 65536 00:25:14.310 } 00:25:14.310 ] 00:25:14.310 }' 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:14.310 20:24:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:14.877 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:14.877 20:24:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:14.877 20:24:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:14.877 20:24:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:25:14.877 20:24:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:14.877 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:14.878 [2024-10-01 20:24:10.012649] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:14.878 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:15.136 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:15.136 "name": "Existed_Raid", 00:25:15.136 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:15.136 "strip_size_kb": 0, 00:25:15.136 "state": "configuring", 00:25:15.136 "raid_level": "raid1", 00:25:15.136 "superblock": false, 00:25:15.136 "num_base_bdevs": 4, 00:25:15.136 "num_base_bdevs_discovered": 2, 00:25:15.136 "num_base_bdevs_operational": 4, 00:25:15.136 "base_bdevs_list": [ 00:25:15.136 { 00:25:15.136 "name": null, 00:25:15.136 "uuid": "daf00519-385f-4951-b4fd-fcf5cf3d8b74", 00:25:15.136 "is_configured": false, 00:25:15.136 "data_offset": 0, 00:25:15.136 "data_size": 65536 00:25:15.136 }, 00:25:15.136 { 00:25:15.136 "name": null, 00:25:15.136 "uuid": "c092e1b1-4405-414f-8dbe-67267435512c", 00:25:15.136 "is_configured": false, 00:25:15.136 "data_offset": 0, 00:25:15.136 "data_size": 65536 00:25:15.136 }, 00:25:15.136 { 00:25:15.136 "name": "BaseBdev3", 00:25:15.136 "uuid": "dce527a3-a7b0-4407-ae6b-9e331428733a", 00:25:15.136 "is_configured": true, 00:25:15.137 "data_offset": 0, 00:25:15.137 "data_size": 65536 00:25:15.137 }, 00:25:15.137 { 00:25:15.137 "name": "BaseBdev4", 00:25:15.137 "uuid": "6f1779e0-4d39-4734-83e1-4c8ac9520198", 00:25:15.137 "is_configured": true, 00:25:15.137 "data_offset": 0, 00:25:15.137 "data_size": 65536 00:25:15.137 } 00:25:15.137 ] 00:25:15.137 }' 00:25:15.137 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:15.137 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:15.396 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:15.396 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:15.396 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:15.396 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:25:15.396 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:15.654 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:25:15.654 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:25:15.654 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:15.654 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:15.654 [2024-10-01 20:24:10.659261] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:25:15.654 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:15.654 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:15.654 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:15.654 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:15.654 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:15.654 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:15.654 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:15.654 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:15.655 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:15.655 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:15.655 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:15.655 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:15.655 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:15.655 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:15.655 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:15.655 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:15.655 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:15.655 "name": "Existed_Raid", 00:25:15.655 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:15.655 "strip_size_kb": 0, 00:25:15.655 "state": "configuring", 00:25:15.655 "raid_level": "raid1", 00:25:15.655 "superblock": false, 00:25:15.655 "num_base_bdevs": 4, 00:25:15.655 "num_base_bdevs_discovered": 3, 00:25:15.655 "num_base_bdevs_operational": 4, 00:25:15.655 "base_bdevs_list": [ 00:25:15.655 { 00:25:15.655 "name": null, 00:25:15.655 "uuid": "daf00519-385f-4951-b4fd-fcf5cf3d8b74", 00:25:15.655 "is_configured": false, 00:25:15.655 "data_offset": 0, 00:25:15.655 "data_size": 65536 00:25:15.655 }, 00:25:15.655 { 00:25:15.655 "name": "BaseBdev2", 00:25:15.655 "uuid": "c092e1b1-4405-414f-8dbe-67267435512c", 00:25:15.655 "is_configured": true, 00:25:15.655 "data_offset": 0, 00:25:15.655 "data_size": 65536 00:25:15.655 }, 00:25:15.655 { 00:25:15.655 "name": "BaseBdev3", 00:25:15.655 "uuid": "dce527a3-a7b0-4407-ae6b-9e331428733a", 00:25:15.655 "is_configured": true, 00:25:15.655 "data_offset": 0, 00:25:15.655 "data_size": 65536 00:25:15.655 }, 00:25:15.655 { 00:25:15.655 "name": "BaseBdev4", 00:25:15.655 "uuid": "6f1779e0-4d39-4734-83e1-4c8ac9520198", 00:25:15.655 "is_configured": true, 00:25:15.655 "data_offset": 0, 00:25:15.655 "data_size": 65536 00:25:15.655 } 00:25:15.655 ] 00:25:15.655 }' 00:25:15.655 20:24:10 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:15.655 20:24:10 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u daf00519-385f-4951-b4fd-fcf5cf3d8b74 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:16.222 [2024-10-01 20:24:11.354119] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:25:16.222 [2024-10-01 20:24:11.354198] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:25:16.222 [2024-10-01 20:24:11.354215] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:25:16.222 [2024-10-01 20:24:11.354590] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:25:16.222 [2024-10-01 20:24:11.354778] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:25:16.222 [2024-10-01 20:24:11.354795] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:25:16.222 [2024-10-01 20:24:11.355214] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:16.222 NewBaseBdev 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:16.222 [ 00:25:16.222 { 00:25:16.222 "name": "NewBaseBdev", 00:25:16.222 "aliases": [ 00:25:16.222 "daf00519-385f-4951-b4fd-fcf5cf3d8b74" 00:25:16.222 ], 00:25:16.222 "product_name": "Malloc disk", 00:25:16.222 "block_size": 512, 00:25:16.222 "num_blocks": 65536, 00:25:16.222 "uuid": "daf00519-385f-4951-b4fd-fcf5cf3d8b74", 00:25:16.222 "assigned_rate_limits": { 00:25:16.222 "rw_ios_per_sec": 0, 00:25:16.222 "rw_mbytes_per_sec": 0, 00:25:16.222 "r_mbytes_per_sec": 0, 00:25:16.222 "w_mbytes_per_sec": 0 00:25:16.222 }, 00:25:16.222 "claimed": true, 00:25:16.222 "claim_type": "exclusive_write", 00:25:16.222 "zoned": false, 00:25:16.222 "supported_io_types": { 00:25:16.222 "read": true, 00:25:16.222 "write": true, 00:25:16.222 "unmap": true, 00:25:16.222 "flush": true, 00:25:16.222 "reset": true, 00:25:16.222 "nvme_admin": false, 00:25:16.222 "nvme_io": false, 00:25:16.222 "nvme_io_md": false, 00:25:16.222 "write_zeroes": true, 00:25:16.222 "zcopy": true, 00:25:16.222 "get_zone_info": false, 00:25:16.222 "zone_management": false, 00:25:16.222 "zone_append": false, 00:25:16.222 "compare": false, 00:25:16.222 "compare_and_write": false, 00:25:16.222 "abort": true, 00:25:16.222 "seek_hole": false, 00:25:16.222 "seek_data": false, 00:25:16.222 "copy": true, 00:25:16.222 "nvme_iov_md": false 00:25:16.222 }, 00:25:16.222 "memory_domains": [ 00:25:16.222 { 00:25:16.222 "dma_device_id": "system", 00:25:16.222 "dma_device_type": 1 00:25:16.222 }, 00:25:16.222 { 00:25:16.222 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:16.222 "dma_device_type": 2 00:25:16.222 } 00:25:16.222 ], 00:25:16.222 "driver_specific": {} 00:25:16.222 } 00:25:16.222 ] 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:16.222 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:16.222 "name": "Existed_Raid", 00:25:16.222 "uuid": "5b5ccdea-7611-49b0-a749-c60ad0e96e42", 00:25:16.222 "strip_size_kb": 0, 00:25:16.222 "state": "online", 00:25:16.222 "raid_level": "raid1", 00:25:16.222 "superblock": false, 00:25:16.222 "num_base_bdevs": 4, 00:25:16.222 "num_base_bdevs_discovered": 4, 00:25:16.222 "num_base_bdevs_operational": 4, 00:25:16.222 "base_bdevs_list": [ 00:25:16.222 { 00:25:16.222 "name": "NewBaseBdev", 00:25:16.222 "uuid": "daf00519-385f-4951-b4fd-fcf5cf3d8b74", 00:25:16.222 "is_configured": true, 00:25:16.222 "data_offset": 0, 00:25:16.222 "data_size": 65536 00:25:16.222 }, 00:25:16.222 { 00:25:16.222 "name": "BaseBdev2", 00:25:16.223 "uuid": "c092e1b1-4405-414f-8dbe-67267435512c", 00:25:16.223 "is_configured": true, 00:25:16.223 "data_offset": 0, 00:25:16.223 "data_size": 65536 00:25:16.223 }, 00:25:16.223 { 00:25:16.223 "name": "BaseBdev3", 00:25:16.223 "uuid": "dce527a3-a7b0-4407-ae6b-9e331428733a", 00:25:16.223 "is_configured": true, 00:25:16.223 "data_offset": 0, 00:25:16.223 "data_size": 65536 00:25:16.223 }, 00:25:16.223 { 00:25:16.223 "name": "BaseBdev4", 00:25:16.223 "uuid": "6f1779e0-4d39-4734-83e1-4c8ac9520198", 00:25:16.223 "is_configured": true, 00:25:16.223 "data_offset": 0, 00:25:16.223 "data_size": 65536 00:25:16.223 } 00:25:16.223 ] 00:25:16.223 }' 00:25:16.223 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:16.223 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:16.789 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:25:16.789 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:25:16.789 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:25:16.789 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:25:16.789 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:25:16.789 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:25:16.789 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:25:16.789 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:25:16.789 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:16.789 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:16.789 [2024-10-01 20:24:11.922808] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:25:16.789 20:24:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:16.789 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:16.789 "name": "Existed_Raid", 00:25:16.789 "aliases": [ 00:25:16.789 "5b5ccdea-7611-49b0-a749-c60ad0e96e42" 00:25:16.789 ], 00:25:16.789 "product_name": "Raid Volume", 00:25:16.789 "block_size": 512, 00:25:16.789 "num_blocks": 65536, 00:25:16.789 "uuid": "5b5ccdea-7611-49b0-a749-c60ad0e96e42", 00:25:16.789 "assigned_rate_limits": { 00:25:16.789 "rw_ios_per_sec": 0, 00:25:16.789 "rw_mbytes_per_sec": 0, 00:25:16.789 "r_mbytes_per_sec": 0, 00:25:16.789 "w_mbytes_per_sec": 0 00:25:16.789 }, 00:25:16.789 "claimed": false, 00:25:16.789 "zoned": false, 00:25:16.789 "supported_io_types": { 00:25:16.789 "read": true, 00:25:16.789 "write": true, 00:25:16.789 "unmap": false, 00:25:16.789 "flush": false, 00:25:16.789 "reset": true, 00:25:16.789 "nvme_admin": false, 00:25:16.789 "nvme_io": false, 00:25:16.789 "nvme_io_md": false, 00:25:16.789 "write_zeroes": true, 00:25:16.789 "zcopy": false, 00:25:16.789 "get_zone_info": false, 00:25:16.789 "zone_management": false, 00:25:16.789 "zone_append": false, 00:25:16.789 "compare": false, 00:25:16.789 "compare_and_write": false, 00:25:16.789 "abort": false, 00:25:16.789 "seek_hole": false, 00:25:16.789 "seek_data": false, 00:25:16.789 "copy": false, 00:25:16.789 "nvme_iov_md": false 00:25:16.789 }, 00:25:16.789 "memory_domains": [ 00:25:16.789 { 00:25:16.789 "dma_device_id": "system", 00:25:16.789 "dma_device_type": 1 00:25:16.789 }, 00:25:16.789 { 00:25:16.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:16.789 "dma_device_type": 2 00:25:16.789 }, 00:25:16.789 { 00:25:16.789 "dma_device_id": "system", 00:25:16.789 "dma_device_type": 1 00:25:16.789 }, 00:25:16.789 { 00:25:16.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:16.789 "dma_device_type": 2 00:25:16.789 }, 00:25:16.789 { 00:25:16.789 "dma_device_id": "system", 00:25:16.789 "dma_device_type": 1 00:25:16.789 }, 00:25:16.789 { 00:25:16.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:16.789 "dma_device_type": 2 00:25:16.790 }, 00:25:16.790 { 00:25:16.790 "dma_device_id": "system", 00:25:16.790 "dma_device_type": 1 00:25:16.790 }, 00:25:16.790 { 00:25:16.790 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:16.790 "dma_device_type": 2 00:25:16.790 } 00:25:16.790 ], 00:25:16.790 "driver_specific": { 00:25:16.790 "raid": { 00:25:16.790 "uuid": "5b5ccdea-7611-49b0-a749-c60ad0e96e42", 00:25:16.790 "strip_size_kb": 0, 00:25:16.790 "state": "online", 00:25:16.790 "raid_level": "raid1", 00:25:16.790 "superblock": false, 00:25:16.790 "num_base_bdevs": 4, 00:25:16.790 "num_base_bdevs_discovered": 4, 00:25:16.790 "num_base_bdevs_operational": 4, 00:25:16.790 "base_bdevs_list": [ 00:25:16.790 { 00:25:16.790 "name": "NewBaseBdev", 00:25:16.790 "uuid": "daf00519-385f-4951-b4fd-fcf5cf3d8b74", 00:25:16.790 "is_configured": true, 00:25:16.790 "data_offset": 0, 00:25:16.790 "data_size": 65536 00:25:16.790 }, 00:25:16.790 { 00:25:16.790 "name": "BaseBdev2", 00:25:16.790 "uuid": "c092e1b1-4405-414f-8dbe-67267435512c", 00:25:16.790 "is_configured": true, 00:25:16.790 "data_offset": 0, 00:25:16.790 "data_size": 65536 00:25:16.790 }, 00:25:16.790 { 00:25:16.790 "name": "BaseBdev3", 00:25:16.790 "uuid": "dce527a3-a7b0-4407-ae6b-9e331428733a", 00:25:16.790 "is_configured": true, 00:25:16.790 "data_offset": 0, 00:25:16.790 "data_size": 65536 00:25:16.790 }, 00:25:16.790 { 00:25:16.790 "name": "BaseBdev4", 00:25:16.790 "uuid": "6f1779e0-4d39-4734-83e1-4c8ac9520198", 00:25:16.790 "is_configured": true, 00:25:16.790 "data_offset": 0, 00:25:16.790 "data_size": 65536 00:25:16.790 } 00:25:16.790 ] 00:25:16.790 } 00:25:16.790 } 00:25:16.790 }' 00:25:16.790 20:24:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:25:16.790 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:25:16.790 BaseBdev2 00:25:16.790 BaseBdev3 00:25:16.790 BaseBdev4' 00:25:16.790 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:17.048 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:17.307 [2024-10-01 20:24:12.302525] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:25:17.307 [2024-10-01 20:24:12.302564] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:25:17.307 [2024-10-01 20:24:12.302677] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:25:17.307 [2024-10-01 20:24:12.303117] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:25:17.307 [2024-10-01 20:24:12.303150] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:25:17.307 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:17.307 20:24:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 74026 00:25:17.307 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 74026 ']' 00:25:17.307 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 74026 00:25:17.307 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:25:17.307 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:25:17.307 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 74026 00:25:17.307 killing process with pid 74026 00:25:17.307 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:25:17.307 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:25:17.307 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 74026' 00:25:17.307 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 74026 00:25:17.307 [2024-10-01 20:24:12.347049] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:25:17.307 20:24:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 74026 00:25:17.565 [2024-10-01 20:24:12.708126] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:25:19.487 00:25:19.487 real 0m13.718s 00:25:19.487 user 0m22.119s 00:25:19.487 sys 0m1.941s 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:25:19.487 ************************************ 00:25:19.487 END TEST raid_state_function_test 00:25:19.487 ************************************ 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:25:19.487 20:24:14 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 4 true 00:25:19.487 20:24:14 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:25:19.487 20:24:14 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:25:19.487 20:24:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:25:19.487 ************************************ 00:25:19.487 START TEST raid_state_function_test_sb 00:25:19.487 ************************************ 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 4 true 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:25:19.487 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:25:19.487 Process raid pid: 74720 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=74720 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 74720' 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 74720 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 74720 ']' 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:19.488 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:25:19.488 20:24:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:19.488 [2024-10-01 20:24:14.726162] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:25:19.488 [2024-10-01 20:24:14.726496] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:25:19.746 [2024-10-01 20:24:14.888627] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:20.004 [2024-10-01 20:24:15.132053] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:25:20.263 [2024-10-01 20:24:15.334625] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:20.263 [2024-10-01 20:24:15.335065] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:20.523 20:24:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:25:20.523 20:24:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:25:20.523 20:24:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:25:20.523 20:24:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:20.523 20:24:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:20.523 [2024-10-01 20:24:15.772862] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:25:20.523 [2024-10-01 20:24:15.773075] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:25:20.523 [2024-10-01 20:24:15.773106] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:25:20.523 [2024-10-01 20:24:15.773127] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:25:20.523 [2024-10-01 20:24:15.773139] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:25:20.523 [2024-10-01 20:24:15.773154] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:25:20.523 [2024-10-01 20:24:15.773175] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:25:20.523 [2024-10-01 20:24:15.773190] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:20.782 "name": "Existed_Raid", 00:25:20.782 "uuid": "c88cc049-09b6-4aa5-956a-927d5ae5eb98", 00:25:20.782 "strip_size_kb": 0, 00:25:20.782 "state": "configuring", 00:25:20.782 "raid_level": "raid1", 00:25:20.782 "superblock": true, 00:25:20.782 "num_base_bdevs": 4, 00:25:20.782 "num_base_bdevs_discovered": 0, 00:25:20.782 "num_base_bdevs_operational": 4, 00:25:20.782 "base_bdevs_list": [ 00:25:20.782 { 00:25:20.782 "name": "BaseBdev1", 00:25:20.782 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:20.782 "is_configured": false, 00:25:20.782 "data_offset": 0, 00:25:20.782 "data_size": 0 00:25:20.782 }, 00:25:20.782 { 00:25:20.782 "name": "BaseBdev2", 00:25:20.782 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:20.782 "is_configured": false, 00:25:20.782 "data_offset": 0, 00:25:20.782 "data_size": 0 00:25:20.782 }, 00:25:20.782 { 00:25:20.782 "name": "BaseBdev3", 00:25:20.782 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:20.782 "is_configured": false, 00:25:20.782 "data_offset": 0, 00:25:20.782 "data_size": 0 00:25:20.782 }, 00:25:20.782 { 00:25:20.782 "name": "BaseBdev4", 00:25:20.782 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:20.782 "is_configured": false, 00:25:20.782 "data_offset": 0, 00:25:20.782 "data_size": 0 00:25:20.782 } 00:25:20.782 ] 00:25:20.782 }' 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:20.782 20:24:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:21.041 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:25:21.041 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:21.041 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:21.041 [2024-10-01 20:24:16.272876] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:25:21.041 [2024-10-01 20:24:16.272932] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:25:21.041 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:21.041 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:25:21.041 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:21.041 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:21.041 [2024-10-01 20:24:16.285009] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:25:21.041 [2024-10-01 20:24:16.285272] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:25:21.041 [2024-10-01 20:24:16.285408] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:25:21.041 [2024-10-01 20:24:16.285549] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:25:21.041 [2024-10-01 20:24:16.285680] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:25:21.041 [2024-10-01 20:24:16.285861] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:25:21.041 [2024-10-01 20:24:16.286003] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:25:21.041 [2024-10-01 20:24:16.286069] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:25:21.041 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:21.041 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:25:21.041 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:21.041 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:21.300 [2024-10-01 20:24:16.332601] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:25:21.300 BaseBdev1 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:21.300 [ 00:25:21.300 { 00:25:21.300 "name": "BaseBdev1", 00:25:21.300 "aliases": [ 00:25:21.300 "84d22d46-14c0-41fc-bdda-164e241e965c" 00:25:21.300 ], 00:25:21.300 "product_name": "Malloc disk", 00:25:21.300 "block_size": 512, 00:25:21.300 "num_blocks": 65536, 00:25:21.300 "uuid": "84d22d46-14c0-41fc-bdda-164e241e965c", 00:25:21.300 "assigned_rate_limits": { 00:25:21.300 "rw_ios_per_sec": 0, 00:25:21.300 "rw_mbytes_per_sec": 0, 00:25:21.300 "r_mbytes_per_sec": 0, 00:25:21.300 "w_mbytes_per_sec": 0 00:25:21.300 }, 00:25:21.300 "claimed": true, 00:25:21.300 "claim_type": "exclusive_write", 00:25:21.300 "zoned": false, 00:25:21.300 "supported_io_types": { 00:25:21.300 "read": true, 00:25:21.300 "write": true, 00:25:21.300 "unmap": true, 00:25:21.300 "flush": true, 00:25:21.300 "reset": true, 00:25:21.300 "nvme_admin": false, 00:25:21.300 "nvme_io": false, 00:25:21.300 "nvme_io_md": false, 00:25:21.300 "write_zeroes": true, 00:25:21.300 "zcopy": true, 00:25:21.300 "get_zone_info": false, 00:25:21.300 "zone_management": false, 00:25:21.300 "zone_append": false, 00:25:21.300 "compare": false, 00:25:21.300 "compare_and_write": false, 00:25:21.300 "abort": true, 00:25:21.300 "seek_hole": false, 00:25:21.300 "seek_data": false, 00:25:21.300 "copy": true, 00:25:21.300 "nvme_iov_md": false 00:25:21.300 }, 00:25:21.300 "memory_domains": [ 00:25:21.300 { 00:25:21.300 "dma_device_id": "system", 00:25:21.300 "dma_device_type": 1 00:25:21.300 }, 00:25:21.300 { 00:25:21.300 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:21.300 "dma_device_type": 2 00:25:21.300 } 00:25:21.300 ], 00:25:21.300 "driver_specific": {} 00:25:21.300 } 00:25:21.300 ] 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:21.300 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:21.300 "name": "Existed_Raid", 00:25:21.300 "uuid": "f2c3d995-e8b5-47f3-b5b4-33c95bf217cf", 00:25:21.300 "strip_size_kb": 0, 00:25:21.300 "state": "configuring", 00:25:21.300 "raid_level": "raid1", 00:25:21.300 "superblock": true, 00:25:21.300 "num_base_bdevs": 4, 00:25:21.300 "num_base_bdevs_discovered": 1, 00:25:21.300 "num_base_bdevs_operational": 4, 00:25:21.300 "base_bdevs_list": [ 00:25:21.300 { 00:25:21.300 "name": "BaseBdev1", 00:25:21.300 "uuid": "84d22d46-14c0-41fc-bdda-164e241e965c", 00:25:21.300 "is_configured": true, 00:25:21.300 "data_offset": 2048, 00:25:21.300 "data_size": 63488 00:25:21.300 }, 00:25:21.300 { 00:25:21.300 "name": "BaseBdev2", 00:25:21.300 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:21.300 "is_configured": false, 00:25:21.300 "data_offset": 0, 00:25:21.300 "data_size": 0 00:25:21.300 }, 00:25:21.300 { 00:25:21.301 "name": "BaseBdev3", 00:25:21.301 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:21.301 "is_configured": false, 00:25:21.301 "data_offset": 0, 00:25:21.301 "data_size": 0 00:25:21.301 }, 00:25:21.301 { 00:25:21.301 "name": "BaseBdev4", 00:25:21.301 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:21.301 "is_configured": false, 00:25:21.301 "data_offset": 0, 00:25:21.301 "data_size": 0 00:25:21.301 } 00:25:21.301 ] 00:25:21.301 }' 00:25:21.301 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:21.301 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:21.869 [2024-10-01 20:24:16.904850] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:25:21.869 [2024-10-01 20:24:16.904923] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:21.869 [2024-10-01 20:24:16.912872] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:25:21.869 [2024-10-01 20:24:16.915541] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:25:21.869 [2024-10-01 20:24:16.915602] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:25:21.869 [2024-10-01 20:24:16.915619] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:25:21.869 [2024-10-01 20:24:16.915639] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:25:21.869 [2024-10-01 20:24:16.915650] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:25:21.869 [2024-10-01 20:24:16.915665] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:21.869 "name": "Existed_Raid", 00:25:21.869 "uuid": "6eee5fde-0f9d-402c-a01d-755185159640", 00:25:21.869 "strip_size_kb": 0, 00:25:21.869 "state": "configuring", 00:25:21.869 "raid_level": "raid1", 00:25:21.869 "superblock": true, 00:25:21.869 "num_base_bdevs": 4, 00:25:21.869 "num_base_bdevs_discovered": 1, 00:25:21.869 "num_base_bdevs_operational": 4, 00:25:21.869 "base_bdevs_list": [ 00:25:21.869 { 00:25:21.869 "name": "BaseBdev1", 00:25:21.869 "uuid": "84d22d46-14c0-41fc-bdda-164e241e965c", 00:25:21.869 "is_configured": true, 00:25:21.869 "data_offset": 2048, 00:25:21.869 "data_size": 63488 00:25:21.869 }, 00:25:21.869 { 00:25:21.869 "name": "BaseBdev2", 00:25:21.869 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:21.869 "is_configured": false, 00:25:21.869 "data_offset": 0, 00:25:21.869 "data_size": 0 00:25:21.869 }, 00:25:21.869 { 00:25:21.869 "name": "BaseBdev3", 00:25:21.869 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:21.869 "is_configured": false, 00:25:21.869 "data_offset": 0, 00:25:21.869 "data_size": 0 00:25:21.869 }, 00:25:21.869 { 00:25:21.869 "name": "BaseBdev4", 00:25:21.869 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:21.869 "is_configured": false, 00:25:21.869 "data_offset": 0, 00:25:21.869 "data_size": 0 00:25:21.869 } 00:25:21.869 ] 00:25:21.869 }' 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:21.869 20:24:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:22.437 [2024-10-01 20:24:17.468822] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:25:22.437 BaseBdev2 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:22.437 [ 00:25:22.437 { 00:25:22.437 "name": "BaseBdev2", 00:25:22.437 "aliases": [ 00:25:22.437 "39bdba35-fdbc-466b-b963-12968af5d333" 00:25:22.437 ], 00:25:22.437 "product_name": "Malloc disk", 00:25:22.437 "block_size": 512, 00:25:22.437 "num_blocks": 65536, 00:25:22.437 "uuid": "39bdba35-fdbc-466b-b963-12968af5d333", 00:25:22.437 "assigned_rate_limits": { 00:25:22.437 "rw_ios_per_sec": 0, 00:25:22.437 "rw_mbytes_per_sec": 0, 00:25:22.437 "r_mbytes_per_sec": 0, 00:25:22.437 "w_mbytes_per_sec": 0 00:25:22.437 }, 00:25:22.437 "claimed": true, 00:25:22.437 "claim_type": "exclusive_write", 00:25:22.437 "zoned": false, 00:25:22.437 "supported_io_types": { 00:25:22.437 "read": true, 00:25:22.437 "write": true, 00:25:22.437 "unmap": true, 00:25:22.437 "flush": true, 00:25:22.437 "reset": true, 00:25:22.437 "nvme_admin": false, 00:25:22.437 "nvme_io": false, 00:25:22.437 "nvme_io_md": false, 00:25:22.437 "write_zeroes": true, 00:25:22.437 "zcopy": true, 00:25:22.437 "get_zone_info": false, 00:25:22.437 "zone_management": false, 00:25:22.437 "zone_append": false, 00:25:22.437 "compare": false, 00:25:22.437 "compare_and_write": false, 00:25:22.437 "abort": true, 00:25:22.437 "seek_hole": false, 00:25:22.437 "seek_data": false, 00:25:22.437 "copy": true, 00:25:22.437 "nvme_iov_md": false 00:25:22.437 }, 00:25:22.437 "memory_domains": [ 00:25:22.437 { 00:25:22.437 "dma_device_id": "system", 00:25:22.437 "dma_device_type": 1 00:25:22.437 }, 00:25:22.437 { 00:25:22.437 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:22.437 "dma_device_type": 2 00:25:22.437 } 00:25:22.437 ], 00:25:22.437 "driver_specific": {} 00:25:22.437 } 00:25:22.437 ] 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:22.437 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:22.438 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:22.438 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:22.438 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:22.438 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:22.438 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:22.438 "name": "Existed_Raid", 00:25:22.438 "uuid": "6eee5fde-0f9d-402c-a01d-755185159640", 00:25:22.438 "strip_size_kb": 0, 00:25:22.438 "state": "configuring", 00:25:22.438 "raid_level": "raid1", 00:25:22.438 "superblock": true, 00:25:22.438 "num_base_bdevs": 4, 00:25:22.438 "num_base_bdevs_discovered": 2, 00:25:22.438 "num_base_bdevs_operational": 4, 00:25:22.438 "base_bdevs_list": [ 00:25:22.438 { 00:25:22.438 "name": "BaseBdev1", 00:25:22.438 "uuid": "84d22d46-14c0-41fc-bdda-164e241e965c", 00:25:22.438 "is_configured": true, 00:25:22.438 "data_offset": 2048, 00:25:22.438 "data_size": 63488 00:25:22.438 }, 00:25:22.438 { 00:25:22.438 "name": "BaseBdev2", 00:25:22.438 "uuid": "39bdba35-fdbc-466b-b963-12968af5d333", 00:25:22.438 "is_configured": true, 00:25:22.438 "data_offset": 2048, 00:25:22.438 "data_size": 63488 00:25:22.438 }, 00:25:22.438 { 00:25:22.438 "name": "BaseBdev3", 00:25:22.438 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:22.438 "is_configured": false, 00:25:22.438 "data_offset": 0, 00:25:22.438 "data_size": 0 00:25:22.438 }, 00:25:22.438 { 00:25:22.438 "name": "BaseBdev4", 00:25:22.438 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:22.438 "is_configured": false, 00:25:22.438 "data_offset": 0, 00:25:22.438 "data_size": 0 00:25:22.438 } 00:25:22.438 ] 00:25:22.438 }' 00:25:22.438 20:24:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:22.438 20:24:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:23.006 [2024-10-01 20:24:18.049258] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:25:23.006 BaseBdev3 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:23.006 [ 00:25:23.006 { 00:25:23.006 "name": "BaseBdev3", 00:25:23.006 "aliases": [ 00:25:23.006 "87128a81-82f0-496d-ad27-7ab2eb757daf" 00:25:23.006 ], 00:25:23.006 "product_name": "Malloc disk", 00:25:23.006 "block_size": 512, 00:25:23.006 "num_blocks": 65536, 00:25:23.006 "uuid": "87128a81-82f0-496d-ad27-7ab2eb757daf", 00:25:23.006 "assigned_rate_limits": { 00:25:23.006 "rw_ios_per_sec": 0, 00:25:23.006 "rw_mbytes_per_sec": 0, 00:25:23.006 "r_mbytes_per_sec": 0, 00:25:23.006 "w_mbytes_per_sec": 0 00:25:23.006 }, 00:25:23.006 "claimed": true, 00:25:23.006 "claim_type": "exclusive_write", 00:25:23.006 "zoned": false, 00:25:23.006 "supported_io_types": { 00:25:23.006 "read": true, 00:25:23.006 "write": true, 00:25:23.006 "unmap": true, 00:25:23.006 "flush": true, 00:25:23.006 "reset": true, 00:25:23.006 "nvme_admin": false, 00:25:23.006 "nvme_io": false, 00:25:23.006 "nvme_io_md": false, 00:25:23.006 "write_zeroes": true, 00:25:23.006 "zcopy": true, 00:25:23.006 "get_zone_info": false, 00:25:23.006 "zone_management": false, 00:25:23.006 "zone_append": false, 00:25:23.006 "compare": false, 00:25:23.006 "compare_and_write": false, 00:25:23.006 "abort": true, 00:25:23.006 "seek_hole": false, 00:25:23.006 "seek_data": false, 00:25:23.006 "copy": true, 00:25:23.006 "nvme_iov_md": false 00:25:23.006 }, 00:25:23.006 "memory_domains": [ 00:25:23.006 { 00:25:23.006 "dma_device_id": "system", 00:25:23.006 "dma_device_type": 1 00:25:23.006 }, 00:25:23.006 { 00:25:23.006 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:23.006 "dma_device_type": 2 00:25:23.006 } 00:25:23.006 ], 00:25:23.006 "driver_specific": {} 00:25:23.006 } 00:25:23.006 ] 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:23.006 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:23.007 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:23.007 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:23.007 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:23.007 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:23.007 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:23.007 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:23.007 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:23.007 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:23.007 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:23.007 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:23.007 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:23.007 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:23.007 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:23.007 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:23.007 "name": "Existed_Raid", 00:25:23.007 "uuid": "6eee5fde-0f9d-402c-a01d-755185159640", 00:25:23.007 "strip_size_kb": 0, 00:25:23.007 "state": "configuring", 00:25:23.007 "raid_level": "raid1", 00:25:23.007 "superblock": true, 00:25:23.007 "num_base_bdevs": 4, 00:25:23.007 "num_base_bdevs_discovered": 3, 00:25:23.007 "num_base_bdevs_operational": 4, 00:25:23.007 "base_bdevs_list": [ 00:25:23.007 { 00:25:23.007 "name": "BaseBdev1", 00:25:23.007 "uuid": "84d22d46-14c0-41fc-bdda-164e241e965c", 00:25:23.007 "is_configured": true, 00:25:23.007 "data_offset": 2048, 00:25:23.007 "data_size": 63488 00:25:23.007 }, 00:25:23.007 { 00:25:23.007 "name": "BaseBdev2", 00:25:23.007 "uuid": "39bdba35-fdbc-466b-b963-12968af5d333", 00:25:23.007 "is_configured": true, 00:25:23.007 "data_offset": 2048, 00:25:23.007 "data_size": 63488 00:25:23.007 }, 00:25:23.007 { 00:25:23.007 "name": "BaseBdev3", 00:25:23.007 "uuid": "87128a81-82f0-496d-ad27-7ab2eb757daf", 00:25:23.007 "is_configured": true, 00:25:23.007 "data_offset": 2048, 00:25:23.007 "data_size": 63488 00:25:23.007 }, 00:25:23.007 { 00:25:23.007 "name": "BaseBdev4", 00:25:23.007 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:23.007 "is_configured": false, 00:25:23.007 "data_offset": 0, 00:25:23.007 "data_size": 0 00:25:23.007 } 00:25:23.007 ] 00:25:23.007 }' 00:25:23.007 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:23.007 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:23.575 [2024-10-01 20:24:18.677399] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:25:23.575 [2024-10-01 20:24:18.677803] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:25:23.575 [2024-10-01 20:24:18.677830] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:25:23.575 BaseBdev4 00:25:23.575 [2024-10-01 20:24:18.678186] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:25:23.575 [2024-10-01 20:24:18.678400] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:25:23.575 [2024-10-01 20:24:18.678425] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:25:23.575 [2024-10-01 20:24:18.678625] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:23.575 [ 00:25:23.575 { 00:25:23.575 "name": "BaseBdev4", 00:25:23.575 "aliases": [ 00:25:23.575 "2164e272-0c57-445a-8bc3-8517b419cf6f" 00:25:23.575 ], 00:25:23.575 "product_name": "Malloc disk", 00:25:23.575 "block_size": 512, 00:25:23.575 "num_blocks": 65536, 00:25:23.575 "uuid": "2164e272-0c57-445a-8bc3-8517b419cf6f", 00:25:23.575 "assigned_rate_limits": { 00:25:23.575 "rw_ios_per_sec": 0, 00:25:23.575 "rw_mbytes_per_sec": 0, 00:25:23.575 "r_mbytes_per_sec": 0, 00:25:23.575 "w_mbytes_per_sec": 0 00:25:23.575 }, 00:25:23.575 "claimed": true, 00:25:23.575 "claim_type": "exclusive_write", 00:25:23.575 "zoned": false, 00:25:23.575 "supported_io_types": { 00:25:23.575 "read": true, 00:25:23.575 "write": true, 00:25:23.575 "unmap": true, 00:25:23.575 "flush": true, 00:25:23.575 "reset": true, 00:25:23.575 "nvme_admin": false, 00:25:23.575 "nvme_io": false, 00:25:23.575 "nvme_io_md": false, 00:25:23.575 "write_zeroes": true, 00:25:23.575 "zcopy": true, 00:25:23.575 "get_zone_info": false, 00:25:23.575 "zone_management": false, 00:25:23.575 "zone_append": false, 00:25:23.575 "compare": false, 00:25:23.575 "compare_and_write": false, 00:25:23.575 "abort": true, 00:25:23.575 "seek_hole": false, 00:25:23.575 "seek_data": false, 00:25:23.575 "copy": true, 00:25:23.575 "nvme_iov_md": false 00:25:23.575 }, 00:25:23.575 "memory_domains": [ 00:25:23.575 { 00:25:23.575 "dma_device_id": "system", 00:25:23.575 "dma_device_type": 1 00:25:23.575 }, 00:25:23.575 { 00:25:23.575 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:23.575 "dma_device_type": 2 00:25:23.575 } 00:25:23.575 ], 00:25:23.575 "driver_specific": {} 00:25:23.575 } 00:25:23.575 ] 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:23.575 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:23.576 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:23.576 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:23.576 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:23.576 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:23.576 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:23.576 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:23.576 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:23.576 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:23.576 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:23.576 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:23.576 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:23.576 "name": "Existed_Raid", 00:25:23.576 "uuid": "6eee5fde-0f9d-402c-a01d-755185159640", 00:25:23.576 "strip_size_kb": 0, 00:25:23.576 "state": "online", 00:25:23.576 "raid_level": "raid1", 00:25:23.576 "superblock": true, 00:25:23.576 "num_base_bdevs": 4, 00:25:23.576 "num_base_bdevs_discovered": 4, 00:25:23.576 "num_base_bdevs_operational": 4, 00:25:23.576 "base_bdevs_list": [ 00:25:23.576 { 00:25:23.576 "name": "BaseBdev1", 00:25:23.576 "uuid": "84d22d46-14c0-41fc-bdda-164e241e965c", 00:25:23.576 "is_configured": true, 00:25:23.576 "data_offset": 2048, 00:25:23.576 "data_size": 63488 00:25:23.576 }, 00:25:23.576 { 00:25:23.576 "name": "BaseBdev2", 00:25:23.576 "uuid": "39bdba35-fdbc-466b-b963-12968af5d333", 00:25:23.576 "is_configured": true, 00:25:23.576 "data_offset": 2048, 00:25:23.576 "data_size": 63488 00:25:23.576 }, 00:25:23.576 { 00:25:23.576 "name": "BaseBdev3", 00:25:23.576 "uuid": "87128a81-82f0-496d-ad27-7ab2eb757daf", 00:25:23.576 "is_configured": true, 00:25:23.576 "data_offset": 2048, 00:25:23.576 "data_size": 63488 00:25:23.576 }, 00:25:23.576 { 00:25:23.576 "name": "BaseBdev4", 00:25:23.576 "uuid": "2164e272-0c57-445a-8bc3-8517b419cf6f", 00:25:23.576 "is_configured": true, 00:25:23.576 "data_offset": 2048, 00:25:23.576 "data_size": 63488 00:25:23.576 } 00:25:23.576 ] 00:25:23.576 }' 00:25:23.576 20:24:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:23.576 20:24:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:24.142 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:25:24.142 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:25:24.142 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:25:24.142 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:25:24.142 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:25:24.142 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:25:24.142 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:25:24.142 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:25:24.142 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:24.142 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:24.142 [2024-10-01 20:24:19.294054] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:25:24.142 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:24.142 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:24.142 "name": "Existed_Raid", 00:25:24.142 "aliases": [ 00:25:24.142 "6eee5fde-0f9d-402c-a01d-755185159640" 00:25:24.142 ], 00:25:24.142 "product_name": "Raid Volume", 00:25:24.142 "block_size": 512, 00:25:24.142 "num_blocks": 63488, 00:25:24.142 "uuid": "6eee5fde-0f9d-402c-a01d-755185159640", 00:25:24.142 "assigned_rate_limits": { 00:25:24.142 "rw_ios_per_sec": 0, 00:25:24.142 "rw_mbytes_per_sec": 0, 00:25:24.142 "r_mbytes_per_sec": 0, 00:25:24.142 "w_mbytes_per_sec": 0 00:25:24.142 }, 00:25:24.142 "claimed": false, 00:25:24.142 "zoned": false, 00:25:24.142 "supported_io_types": { 00:25:24.142 "read": true, 00:25:24.142 "write": true, 00:25:24.142 "unmap": false, 00:25:24.142 "flush": false, 00:25:24.142 "reset": true, 00:25:24.142 "nvme_admin": false, 00:25:24.142 "nvme_io": false, 00:25:24.142 "nvme_io_md": false, 00:25:24.142 "write_zeroes": true, 00:25:24.142 "zcopy": false, 00:25:24.142 "get_zone_info": false, 00:25:24.142 "zone_management": false, 00:25:24.142 "zone_append": false, 00:25:24.142 "compare": false, 00:25:24.142 "compare_and_write": false, 00:25:24.142 "abort": false, 00:25:24.142 "seek_hole": false, 00:25:24.142 "seek_data": false, 00:25:24.142 "copy": false, 00:25:24.142 "nvme_iov_md": false 00:25:24.142 }, 00:25:24.142 "memory_domains": [ 00:25:24.142 { 00:25:24.142 "dma_device_id": "system", 00:25:24.142 "dma_device_type": 1 00:25:24.142 }, 00:25:24.142 { 00:25:24.142 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:24.142 "dma_device_type": 2 00:25:24.142 }, 00:25:24.142 { 00:25:24.142 "dma_device_id": "system", 00:25:24.142 "dma_device_type": 1 00:25:24.142 }, 00:25:24.142 { 00:25:24.142 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:24.142 "dma_device_type": 2 00:25:24.142 }, 00:25:24.142 { 00:25:24.142 "dma_device_id": "system", 00:25:24.142 "dma_device_type": 1 00:25:24.142 }, 00:25:24.142 { 00:25:24.142 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:24.142 "dma_device_type": 2 00:25:24.142 }, 00:25:24.142 { 00:25:24.142 "dma_device_id": "system", 00:25:24.142 "dma_device_type": 1 00:25:24.142 }, 00:25:24.142 { 00:25:24.142 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:24.142 "dma_device_type": 2 00:25:24.142 } 00:25:24.142 ], 00:25:24.142 "driver_specific": { 00:25:24.142 "raid": { 00:25:24.142 "uuid": "6eee5fde-0f9d-402c-a01d-755185159640", 00:25:24.142 "strip_size_kb": 0, 00:25:24.142 "state": "online", 00:25:24.142 "raid_level": "raid1", 00:25:24.142 "superblock": true, 00:25:24.142 "num_base_bdevs": 4, 00:25:24.142 "num_base_bdevs_discovered": 4, 00:25:24.142 "num_base_bdevs_operational": 4, 00:25:24.142 "base_bdevs_list": [ 00:25:24.142 { 00:25:24.142 "name": "BaseBdev1", 00:25:24.142 "uuid": "84d22d46-14c0-41fc-bdda-164e241e965c", 00:25:24.142 "is_configured": true, 00:25:24.142 "data_offset": 2048, 00:25:24.142 "data_size": 63488 00:25:24.142 }, 00:25:24.142 { 00:25:24.142 "name": "BaseBdev2", 00:25:24.142 "uuid": "39bdba35-fdbc-466b-b963-12968af5d333", 00:25:24.142 "is_configured": true, 00:25:24.142 "data_offset": 2048, 00:25:24.142 "data_size": 63488 00:25:24.142 }, 00:25:24.142 { 00:25:24.142 "name": "BaseBdev3", 00:25:24.142 "uuid": "87128a81-82f0-496d-ad27-7ab2eb757daf", 00:25:24.142 "is_configured": true, 00:25:24.142 "data_offset": 2048, 00:25:24.142 "data_size": 63488 00:25:24.142 }, 00:25:24.142 { 00:25:24.142 "name": "BaseBdev4", 00:25:24.142 "uuid": "2164e272-0c57-445a-8bc3-8517b419cf6f", 00:25:24.142 "is_configured": true, 00:25:24.142 "data_offset": 2048, 00:25:24.142 "data_size": 63488 00:25:24.142 } 00:25:24.142 ] 00:25:24.142 } 00:25:24.142 } 00:25:24.142 }' 00:25:24.142 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:25:24.142 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:25:24.142 BaseBdev2 00:25:24.142 BaseBdev3 00:25:24.142 BaseBdev4' 00:25:24.142 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:24.402 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:24.402 [2024-10-01 20:24:19.653871] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:24.662 "name": "Existed_Raid", 00:25:24.662 "uuid": "6eee5fde-0f9d-402c-a01d-755185159640", 00:25:24.662 "strip_size_kb": 0, 00:25:24.662 "state": "online", 00:25:24.662 "raid_level": "raid1", 00:25:24.662 "superblock": true, 00:25:24.662 "num_base_bdevs": 4, 00:25:24.662 "num_base_bdevs_discovered": 3, 00:25:24.662 "num_base_bdevs_operational": 3, 00:25:24.662 "base_bdevs_list": [ 00:25:24.662 { 00:25:24.662 "name": null, 00:25:24.662 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:24.662 "is_configured": false, 00:25:24.662 "data_offset": 0, 00:25:24.662 "data_size": 63488 00:25:24.662 }, 00:25:24.662 { 00:25:24.662 "name": "BaseBdev2", 00:25:24.662 "uuid": "39bdba35-fdbc-466b-b963-12968af5d333", 00:25:24.662 "is_configured": true, 00:25:24.662 "data_offset": 2048, 00:25:24.662 "data_size": 63488 00:25:24.662 }, 00:25:24.662 { 00:25:24.662 "name": "BaseBdev3", 00:25:24.662 "uuid": "87128a81-82f0-496d-ad27-7ab2eb757daf", 00:25:24.662 "is_configured": true, 00:25:24.662 "data_offset": 2048, 00:25:24.662 "data_size": 63488 00:25:24.662 }, 00:25:24.662 { 00:25:24.662 "name": "BaseBdev4", 00:25:24.662 "uuid": "2164e272-0c57-445a-8bc3-8517b419cf6f", 00:25:24.662 "is_configured": true, 00:25:24.662 "data_offset": 2048, 00:25:24.662 "data_size": 63488 00:25:24.662 } 00:25:24.662 ] 00:25:24.662 }' 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:24.662 20:24:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.230 [2024-10-01 20:24:20.334385] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:25.230 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.489 [2024-10-01 20:24:20.483914] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.489 [2024-10-01 20:24:20.627359] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:25:25.489 [2024-10-01 20:24:20.627517] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:25:25.489 [2024-10-01 20:24:20.715416] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:25:25.489 [2024-10-01 20:24:20.715519] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:25:25.489 [2024-10-01 20:24:20.715542] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.489 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.748 BaseBdev2 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:25:25.748 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.749 [ 00:25:25.749 { 00:25:25.749 "name": "BaseBdev2", 00:25:25.749 "aliases": [ 00:25:25.749 "150ef83d-f5f6-401b-88b5-e4bc90380427" 00:25:25.749 ], 00:25:25.749 "product_name": "Malloc disk", 00:25:25.749 "block_size": 512, 00:25:25.749 "num_blocks": 65536, 00:25:25.749 "uuid": "150ef83d-f5f6-401b-88b5-e4bc90380427", 00:25:25.749 "assigned_rate_limits": { 00:25:25.749 "rw_ios_per_sec": 0, 00:25:25.749 "rw_mbytes_per_sec": 0, 00:25:25.749 "r_mbytes_per_sec": 0, 00:25:25.749 "w_mbytes_per_sec": 0 00:25:25.749 }, 00:25:25.749 "claimed": false, 00:25:25.749 "zoned": false, 00:25:25.749 "supported_io_types": { 00:25:25.749 "read": true, 00:25:25.749 "write": true, 00:25:25.749 "unmap": true, 00:25:25.749 "flush": true, 00:25:25.749 "reset": true, 00:25:25.749 "nvme_admin": false, 00:25:25.749 "nvme_io": false, 00:25:25.749 "nvme_io_md": false, 00:25:25.749 "write_zeroes": true, 00:25:25.749 "zcopy": true, 00:25:25.749 "get_zone_info": false, 00:25:25.749 "zone_management": false, 00:25:25.749 "zone_append": false, 00:25:25.749 "compare": false, 00:25:25.749 "compare_and_write": false, 00:25:25.749 "abort": true, 00:25:25.749 "seek_hole": false, 00:25:25.749 "seek_data": false, 00:25:25.749 "copy": true, 00:25:25.749 "nvme_iov_md": false 00:25:25.749 }, 00:25:25.749 "memory_domains": [ 00:25:25.749 { 00:25:25.749 "dma_device_id": "system", 00:25:25.749 "dma_device_type": 1 00:25:25.749 }, 00:25:25.749 { 00:25:25.749 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:25.749 "dma_device_type": 2 00:25:25.749 } 00:25:25.749 ], 00:25:25.749 "driver_specific": {} 00:25:25.749 } 00:25:25.749 ] 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.749 BaseBdev3 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.749 [ 00:25:25.749 { 00:25:25.749 "name": "BaseBdev3", 00:25:25.749 "aliases": [ 00:25:25.749 "d146c0d7-9bad-472b-9811-73868faa7ada" 00:25:25.749 ], 00:25:25.749 "product_name": "Malloc disk", 00:25:25.749 "block_size": 512, 00:25:25.749 "num_blocks": 65536, 00:25:25.749 "uuid": "d146c0d7-9bad-472b-9811-73868faa7ada", 00:25:25.749 "assigned_rate_limits": { 00:25:25.749 "rw_ios_per_sec": 0, 00:25:25.749 "rw_mbytes_per_sec": 0, 00:25:25.749 "r_mbytes_per_sec": 0, 00:25:25.749 "w_mbytes_per_sec": 0 00:25:25.749 }, 00:25:25.749 "claimed": false, 00:25:25.749 "zoned": false, 00:25:25.749 "supported_io_types": { 00:25:25.749 "read": true, 00:25:25.749 "write": true, 00:25:25.749 "unmap": true, 00:25:25.749 "flush": true, 00:25:25.749 "reset": true, 00:25:25.749 "nvme_admin": false, 00:25:25.749 "nvme_io": false, 00:25:25.749 "nvme_io_md": false, 00:25:25.749 "write_zeroes": true, 00:25:25.749 "zcopy": true, 00:25:25.749 "get_zone_info": false, 00:25:25.749 "zone_management": false, 00:25:25.749 "zone_append": false, 00:25:25.749 "compare": false, 00:25:25.749 "compare_and_write": false, 00:25:25.749 "abort": true, 00:25:25.749 "seek_hole": false, 00:25:25.749 "seek_data": false, 00:25:25.749 "copy": true, 00:25:25.749 "nvme_iov_md": false 00:25:25.749 }, 00:25:25.749 "memory_domains": [ 00:25:25.749 { 00:25:25.749 "dma_device_id": "system", 00:25:25.749 "dma_device_type": 1 00:25:25.749 }, 00:25:25.749 { 00:25:25.749 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:25.749 "dma_device_type": 2 00:25:25.749 } 00:25:25.749 ], 00:25:25.749 "driver_specific": {} 00:25:25.749 } 00:25:25.749 ] 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.749 BaseBdev4 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:25.749 20:24:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:25.749 [ 00:25:25.749 { 00:25:25.749 "name": "BaseBdev4", 00:25:25.749 "aliases": [ 00:25:25.749 "0473aaa1-4d55-4376-9527-77e6a5b1da3f" 00:25:25.749 ], 00:25:25.749 "product_name": "Malloc disk", 00:25:25.749 "block_size": 512, 00:25:25.749 "num_blocks": 65536, 00:25:25.749 "uuid": "0473aaa1-4d55-4376-9527-77e6a5b1da3f", 00:25:25.749 "assigned_rate_limits": { 00:25:25.749 "rw_ios_per_sec": 0, 00:25:25.749 "rw_mbytes_per_sec": 0, 00:25:25.749 "r_mbytes_per_sec": 0, 00:25:25.749 "w_mbytes_per_sec": 0 00:25:25.749 }, 00:25:25.749 "claimed": false, 00:25:25.749 "zoned": false, 00:25:25.749 "supported_io_types": { 00:25:25.749 "read": true, 00:25:25.749 "write": true, 00:25:25.749 "unmap": true, 00:25:25.749 "flush": true, 00:25:25.749 "reset": true, 00:25:25.749 "nvme_admin": false, 00:25:25.749 "nvme_io": false, 00:25:25.749 "nvme_io_md": false, 00:25:25.749 "write_zeroes": true, 00:25:25.749 "zcopy": true, 00:25:25.749 "get_zone_info": false, 00:25:25.749 "zone_management": false, 00:25:25.749 "zone_append": false, 00:25:25.749 "compare": false, 00:25:25.749 "compare_and_write": false, 00:25:25.749 "abort": true, 00:25:25.749 "seek_hole": false, 00:25:25.749 "seek_data": false, 00:25:25.749 "copy": true, 00:25:25.749 "nvme_iov_md": false 00:25:25.749 }, 00:25:25.749 "memory_domains": [ 00:25:26.008 { 00:25:26.008 "dma_device_id": "system", 00:25:26.008 "dma_device_type": 1 00:25:26.008 }, 00:25:26.008 { 00:25:26.008 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:26.008 "dma_device_type": 2 00:25:26.008 } 00:25:26.008 ], 00:25:26.008 "driver_specific": {} 00:25:26.008 } 00:25:26.008 ] 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:26.008 [2024-10-01 20:24:21.010842] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:25:26.008 [2024-10-01 20:24:21.010924] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:25:26.008 [2024-10-01 20:24:21.010954] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:25:26.008 [2024-10-01 20:24:21.013646] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:25:26.008 [2024-10-01 20:24:21.013774] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:26.008 "name": "Existed_Raid", 00:25:26.008 "uuid": "da79bddb-7716-4558-acc0-00cd9de7521f", 00:25:26.008 "strip_size_kb": 0, 00:25:26.008 "state": "configuring", 00:25:26.008 "raid_level": "raid1", 00:25:26.008 "superblock": true, 00:25:26.008 "num_base_bdevs": 4, 00:25:26.008 "num_base_bdevs_discovered": 3, 00:25:26.008 "num_base_bdevs_operational": 4, 00:25:26.008 "base_bdevs_list": [ 00:25:26.008 { 00:25:26.008 "name": "BaseBdev1", 00:25:26.008 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:26.008 "is_configured": false, 00:25:26.008 "data_offset": 0, 00:25:26.008 "data_size": 0 00:25:26.008 }, 00:25:26.008 { 00:25:26.008 "name": "BaseBdev2", 00:25:26.008 "uuid": "150ef83d-f5f6-401b-88b5-e4bc90380427", 00:25:26.008 "is_configured": true, 00:25:26.008 "data_offset": 2048, 00:25:26.008 "data_size": 63488 00:25:26.008 }, 00:25:26.008 { 00:25:26.008 "name": "BaseBdev3", 00:25:26.008 "uuid": "d146c0d7-9bad-472b-9811-73868faa7ada", 00:25:26.008 "is_configured": true, 00:25:26.008 "data_offset": 2048, 00:25:26.008 "data_size": 63488 00:25:26.008 }, 00:25:26.008 { 00:25:26.008 "name": "BaseBdev4", 00:25:26.008 "uuid": "0473aaa1-4d55-4376-9527-77e6a5b1da3f", 00:25:26.008 "is_configured": true, 00:25:26.008 "data_offset": 2048, 00:25:26.008 "data_size": 63488 00:25:26.008 } 00:25:26.008 ] 00:25:26.008 }' 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:26.008 20:24:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:26.576 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:25:26.576 20:24:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:26.576 20:24:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:26.576 [2024-10-01 20:24:21.551049] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:25:26.576 20:24:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:26.576 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:26.576 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:26.576 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:26.576 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:26.576 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:26.576 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:26.576 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:26.576 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:26.576 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:26.576 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:26.576 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:26.576 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:26.577 20:24:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:26.577 20:24:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:26.577 20:24:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:26.577 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:26.577 "name": "Existed_Raid", 00:25:26.577 "uuid": "da79bddb-7716-4558-acc0-00cd9de7521f", 00:25:26.577 "strip_size_kb": 0, 00:25:26.577 "state": "configuring", 00:25:26.577 "raid_level": "raid1", 00:25:26.577 "superblock": true, 00:25:26.577 "num_base_bdevs": 4, 00:25:26.577 "num_base_bdevs_discovered": 2, 00:25:26.577 "num_base_bdevs_operational": 4, 00:25:26.577 "base_bdevs_list": [ 00:25:26.577 { 00:25:26.577 "name": "BaseBdev1", 00:25:26.577 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:26.577 "is_configured": false, 00:25:26.577 "data_offset": 0, 00:25:26.577 "data_size": 0 00:25:26.577 }, 00:25:26.577 { 00:25:26.577 "name": null, 00:25:26.577 "uuid": "150ef83d-f5f6-401b-88b5-e4bc90380427", 00:25:26.577 "is_configured": false, 00:25:26.577 "data_offset": 0, 00:25:26.577 "data_size": 63488 00:25:26.577 }, 00:25:26.577 { 00:25:26.577 "name": "BaseBdev3", 00:25:26.577 "uuid": "d146c0d7-9bad-472b-9811-73868faa7ada", 00:25:26.577 "is_configured": true, 00:25:26.577 "data_offset": 2048, 00:25:26.577 "data_size": 63488 00:25:26.577 }, 00:25:26.577 { 00:25:26.577 "name": "BaseBdev4", 00:25:26.577 "uuid": "0473aaa1-4d55-4376-9527-77e6a5b1da3f", 00:25:26.577 "is_configured": true, 00:25:26.577 "data_offset": 2048, 00:25:26.577 "data_size": 63488 00:25:26.577 } 00:25:26.577 ] 00:25:26.577 }' 00:25:26.577 20:24:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:26.577 20:24:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:26.841 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:26.841 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:26.841 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:26.841 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:25:26.841 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:27.101 [2024-10-01 20:24:22.161557] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:25:27.101 BaseBdev1 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:27.101 [ 00:25:27.101 { 00:25:27.101 "name": "BaseBdev1", 00:25:27.101 "aliases": [ 00:25:27.101 "1400b88a-2a60-4e39-99f1-c2effa87de04" 00:25:27.101 ], 00:25:27.101 "product_name": "Malloc disk", 00:25:27.101 "block_size": 512, 00:25:27.101 "num_blocks": 65536, 00:25:27.101 "uuid": "1400b88a-2a60-4e39-99f1-c2effa87de04", 00:25:27.101 "assigned_rate_limits": { 00:25:27.101 "rw_ios_per_sec": 0, 00:25:27.101 "rw_mbytes_per_sec": 0, 00:25:27.101 "r_mbytes_per_sec": 0, 00:25:27.101 "w_mbytes_per_sec": 0 00:25:27.101 }, 00:25:27.101 "claimed": true, 00:25:27.101 "claim_type": "exclusive_write", 00:25:27.101 "zoned": false, 00:25:27.101 "supported_io_types": { 00:25:27.101 "read": true, 00:25:27.101 "write": true, 00:25:27.101 "unmap": true, 00:25:27.101 "flush": true, 00:25:27.101 "reset": true, 00:25:27.101 "nvme_admin": false, 00:25:27.101 "nvme_io": false, 00:25:27.101 "nvme_io_md": false, 00:25:27.101 "write_zeroes": true, 00:25:27.101 "zcopy": true, 00:25:27.101 "get_zone_info": false, 00:25:27.101 "zone_management": false, 00:25:27.101 "zone_append": false, 00:25:27.101 "compare": false, 00:25:27.101 "compare_and_write": false, 00:25:27.101 "abort": true, 00:25:27.101 "seek_hole": false, 00:25:27.101 "seek_data": false, 00:25:27.101 "copy": true, 00:25:27.101 "nvme_iov_md": false 00:25:27.101 }, 00:25:27.101 "memory_domains": [ 00:25:27.101 { 00:25:27.101 "dma_device_id": "system", 00:25:27.101 "dma_device_type": 1 00:25:27.101 }, 00:25:27.101 { 00:25:27.101 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:27.101 "dma_device_type": 2 00:25:27.101 } 00:25:27.101 ], 00:25:27.101 "driver_specific": {} 00:25:27.101 } 00:25:27.101 ] 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:27.101 "name": "Existed_Raid", 00:25:27.101 "uuid": "da79bddb-7716-4558-acc0-00cd9de7521f", 00:25:27.101 "strip_size_kb": 0, 00:25:27.101 "state": "configuring", 00:25:27.101 "raid_level": "raid1", 00:25:27.101 "superblock": true, 00:25:27.101 "num_base_bdevs": 4, 00:25:27.101 "num_base_bdevs_discovered": 3, 00:25:27.101 "num_base_bdevs_operational": 4, 00:25:27.101 "base_bdevs_list": [ 00:25:27.101 { 00:25:27.101 "name": "BaseBdev1", 00:25:27.101 "uuid": "1400b88a-2a60-4e39-99f1-c2effa87de04", 00:25:27.101 "is_configured": true, 00:25:27.101 "data_offset": 2048, 00:25:27.101 "data_size": 63488 00:25:27.101 }, 00:25:27.101 { 00:25:27.101 "name": null, 00:25:27.101 "uuid": "150ef83d-f5f6-401b-88b5-e4bc90380427", 00:25:27.101 "is_configured": false, 00:25:27.101 "data_offset": 0, 00:25:27.101 "data_size": 63488 00:25:27.101 }, 00:25:27.101 { 00:25:27.101 "name": "BaseBdev3", 00:25:27.101 "uuid": "d146c0d7-9bad-472b-9811-73868faa7ada", 00:25:27.101 "is_configured": true, 00:25:27.101 "data_offset": 2048, 00:25:27.101 "data_size": 63488 00:25:27.101 }, 00:25:27.101 { 00:25:27.101 "name": "BaseBdev4", 00:25:27.101 "uuid": "0473aaa1-4d55-4376-9527-77e6a5b1da3f", 00:25:27.101 "is_configured": true, 00:25:27.101 "data_offset": 2048, 00:25:27.101 "data_size": 63488 00:25:27.101 } 00:25:27.101 ] 00:25:27.101 }' 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:27.101 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:27.669 [2024-10-01 20:24:22.777867] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:27.669 "name": "Existed_Raid", 00:25:27.669 "uuid": "da79bddb-7716-4558-acc0-00cd9de7521f", 00:25:27.669 "strip_size_kb": 0, 00:25:27.669 "state": "configuring", 00:25:27.669 "raid_level": "raid1", 00:25:27.669 "superblock": true, 00:25:27.669 "num_base_bdevs": 4, 00:25:27.669 "num_base_bdevs_discovered": 2, 00:25:27.669 "num_base_bdevs_operational": 4, 00:25:27.669 "base_bdevs_list": [ 00:25:27.669 { 00:25:27.669 "name": "BaseBdev1", 00:25:27.669 "uuid": "1400b88a-2a60-4e39-99f1-c2effa87de04", 00:25:27.669 "is_configured": true, 00:25:27.669 "data_offset": 2048, 00:25:27.669 "data_size": 63488 00:25:27.669 }, 00:25:27.669 { 00:25:27.669 "name": null, 00:25:27.669 "uuid": "150ef83d-f5f6-401b-88b5-e4bc90380427", 00:25:27.669 "is_configured": false, 00:25:27.669 "data_offset": 0, 00:25:27.669 "data_size": 63488 00:25:27.669 }, 00:25:27.669 { 00:25:27.669 "name": null, 00:25:27.669 "uuid": "d146c0d7-9bad-472b-9811-73868faa7ada", 00:25:27.669 "is_configured": false, 00:25:27.669 "data_offset": 0, 00:25:27.669 "data_size": 63488 00:25:27.669 }, 00:25:27.669 { 00:25:27.669 "name": "BaseBdev4", 00:25:27.669 "uuid": "0473aaa1-4d55-4376-9527-77e6a5b1da3f", 00:25:27.669 "is_configured": true, 00:25:27.669 "data_offset": 2048, 00:25:27.669 "data_size": 63488 00:25:27.669 } 00:25:27.669 ] 00:25:27.669 }' 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:27.669 20:24:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:28.237 [2024-10-01 20:24:23.342011] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:28.237 "name": "Existed_Raid", 00:25:28.237 "uuid": "da79bddb-7716-4558-acc0-00cd9de7521f", 00:25:28.237 "strip_size_kb": 0, 00:25:28.237 "state": "configuring", 00:25:28.237 "raid_level": "raid1", 00:25:28.237 "superblock": true, 00:25:28.237 "num_base_bdevs": 4, 00:25:28.237 "num_base_bdevs_discovered": 3, 00:25:28.237 "num_base_bdevs_operational": 4, 00:25:28.237 "base_bdevs_list": [ 00:25:28.237 { 00:25:28.237 "name": "BaseBdev1", 00:25:28.237 "uuid": "1400b88a-2a60-4e39-99f1-c2effa87de04", 00:25:28.237 "is_configured": true, 00:25:28.237 "data_offset": 2048, 00:25:28.237 "data_size": 63488 00:25:28.237 }, 00:25:28.237 { 00:25:28.237 "name": null, 00:25:28.237 "uuid": "150ef83d-f5f6-401b-88b5-e4bc90380427", 00:25:28.237 "is_configured": false, 00:25:28.237 "data_offset": 0, 00:25:28.237 "data_size": 63488 00:25:28.237 }, 00:25:28.237 { 00:25:28.237 "name": "BaseBdev3", 00:25:28.237 "uuid": "d146c0d7-9bad-472b-9811-73868faa7ada", 00:25:28.237 "is_configured": true, 00:25:28.237 "data_offset": 2048, 00:25:28.237 "data_size": 63488 00:25:28.237 }, 00:25:28.237 { 00:25:28.237 "name": "BaseBdev4", 00:25:28.237 "uuid": "0473aaa1-4d55-4376-9527-77e6a5b1da3f", 00:25:28.237 "is_configured": true, 00:25:28.237 "data_offset": 2048, 00:25:28.237 "data_size": 63488 00:25:28.237 } 00:25:28.237 ] 00:25:28.237 }' 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:28.237 20:24:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:28.805 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:28.805 20:24:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:28.805 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:25:28.805 20:24:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:28.805 20:24:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:28.805 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:25:28.805 20:24:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:25:28.805 20:24:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:28.805 20:24:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:28.805 [2024-10-01 20:24:23.926217] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:25:28.805 20:24:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:28.805 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:28.805 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:28.805 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:28.805 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:28.805 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:28.805 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:28.805 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:28.805 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:28.805 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:28.805 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:28.805 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:28.805 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:28.805 20:24:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:28.805 20:24:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:28.805 20:24:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:29.064 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:29.064 "name": "Existed_Raid", 00:25:29.064 "uuid": "da79bddb-7716-4558-acc0-00cd9de7521f", 00:25:29.064 "strip_size_kb": 0, 00:25:29.064 "state": "configuring", 00:25:29.064 "raid_level": "raid1", 00:25:29.064 "superblock": true, 00:25:29.064 "num_base_bdevs": 4, 00:25:29.064 "num_base_bdevs_discovered": 2, 00:25:29.064 "num_base_bdevs_operational": 4, 00:25:29.064 "base_bdevs_list": [ 00:25:29.064 { 00:25:29.064 "name": null, 00:25:29.064 "uuid": "1400b88a-2a60-4e39-99f1-c2effa87de04", 00:25:29.064 "is_configured": false, 00:25:29.064 "data_offset": 0, 00:25:29.064 "data_size": 63488 00:25:29.064 }, 00:25:29.064 { 00:25:29.064 "name": null, 00:25:29.064 "uuid": "150ef83d-f5f6-401b-88b5-e4bc90380427", 00:25:29.064 "is_configured": false, 00:25:29.064 "data_offset": 0, 00:25:29.064 "data_size": 63488 00:25:29.064 }, 00:25:29.064 { 00:25:29.064 "name": "BaseBdev3", 00:25:29.064 "uuid": "d146c0d7-9bad-472b-9811-73868faa7ada", 00:25:29.064 "is_configured": true, 00:25:29.064 "data_offset": 2048, 00:25:29.064 "data_size": 63488 00:25:29.064 }, 00:25:29.064 { 00:25:29.064 "name": "BaseBdev4", 00:25:29.064 "uuid": "0473aaa1-4d55-4376-9527-77e6a5b1da3f", 00:25:29.064 "is_configured": true, 00:25:29.064 "data_offset": 2048, 00:25:29.064 "data_size": 63488 00:25:29.064 } 00:25:29.064 ] 00:25:29.064 }' 00:25:29.064 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:29.064 20:24:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:29.323 [2024-10-01 20:24:24.566558] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:29.323 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:29.583 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:29.583 20:24:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:29.583 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:29.583 20:24:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:29.583 20:24:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:29.583 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:29.583 "name": "Existed_Raid", 00:25:29.583 "uuid": "da79bddb-7716-4558-acc0-00cd9de7521f", 00:25:29.583 "strip_size_kb": 0, 00:25:29.583 "state": "configuring", 00:25:29.583 "raid_level": "raid1", 00:25:29.583 "superblock": true, 00:25:29.583 "num_base_bdevs": 4, 00:25:29.583 "num_base_bdevs_discovered": 3, 00:25:29.583 "num_base_bdevs_operational": 4, 00:25:29.583 "base_bdevs_list": [ 00:25:29.583 { 00:25:29.583 "name": null, 00:25:29.583 "uuid": "1400b88a-2a60-4e39-99f1-c2effa87de04", 00:25:29.583 "is_configured": false, 00:25:29.583 "data_offset": 0, 00:25:29.583 "data_size": 63488 00:25:29.583 }, 00:25:29.583 { 00:25:29.583 "name": "BaseBdev2", 00:25:29.583 "uuid": "150ef83d-f5f6-401b-88b5-e4bc90380427", 00:25:29.583 "is_configured": true, 00:25:29.583 "data_offset": 2048, 00:25:29.583 "data_size": 63488 00:25:29.583 }, 00:25:29.583 { 00:25:29.583 "name": "BaseBdev3", 00:25:29.583 "uuid": "d146c0d7-9bad-472b-9811-73868faa7ada", 00:25:29.583 "is_configured": true, 00:25:29.583 "data_offset": 2048, 00:25:29.583 "data_size": 63488 00:25:29.583 }, 00:25:29.583 { 00:25:29.583 "name": "BaseBdev4", 00:25:29.583 "uuid": "0473aaa1-4d55-4376-9527-77e6a5b1da3f", 00:25:29.583 "is_configured": true, 00:25:29.583 "data_offset": 2048, 00:25:29.583 "data_size": 63488 00:25:29.583 } 00:25:29.583 ] 00:25:29.583 }' 00:25:29.583 20:24:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:29.583 20:24:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:30.152 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:30.152 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.152 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:30.152 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:25:30.152 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.152 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:25:30.152 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:30.152 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.152 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:30.152 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:25:30.152 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 1400b88a-2a60-4e39-99f1-c2effa87de04 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:30.153 [2024-10-01 20:24:25.257708] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:25:30.153 [2024-10-01 20:24:25.258051] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:25:30.153 [2024-10-01 20:24:25.258077] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:25:30.153 [2024-10-01 20:24:25.258438] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:25:30.153 [2024-10-01 20:24:25.258658] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:25:30.153 [2024-10-01 20:24:25.258682] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:25:30.153 [2024-10-01 20:24:25.258885] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:30.153 NewBaseBdev 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:30.153 [ 00:25:30.153 { 00:25:30.153 "name": "NewBaseBdev", 00:25:30.153 "aliases": [ 00:25:30.153 "1400b88a-2a60-4e39-99f1-c2effa87de04" 00:25:30.153 ], 00:25:30.153 "product_name": "Malloc disk", 00:25:30.153 "block_size": 512, 00:25:30.153 "num_blocks": 65536, 00:25:30.153 "uuid": "1400b88a-2a60-4e39-99f1-c2effa87de04", 00:25:30.153 "assigned_rate_limits": { 00:25:30.153 "rw_ios_per_sec": 0, 00:25:30.153 "rw_mbytes_per_sec": 0, 00:25:30.153 "r_mbytes_per_sec": 0, 00:25:30.153 "w_mbytes_per_sec": 0 00:25:30.153 }, 00:25:30.153 "claimed": true, 00:25:30.153 "claim_type": "exclusive_write", 00:25:30.153 "zoned": false, 00:25:30.153 "supported_io_types": { 00:25:30.153 "read": true, 00:25:30.153 "write": true, 00:25:30.153 "unmap": true, 00:25:30.153 "flush": true, 00:25:30.153 "reset": true, 00:25:30.153 "nvme_admin": false, 00:25:30.153 "nvme_io": false, 00:25:30.153 "nvme_io_md": false, 00:25:30.153 "write_zeroes": true, 00:25:30.153 "zcopy": true, 00:25:30.153 "get_zone_info": false, 00:25:30.153 "zone_management": false, 00:25:30.153 "zone_append": false, 00:25:30.153 "compare": false, 00:25:30.153 "compare_and_write": false, 00:25:30.153 "abort": true, 00:25:30.153 "seek_hole": false, 00:25:30.153 "seek_data": false, 00:25:30.153 "copy": true, 00:25:30.153 "nvme_iov_md": false 00:25:30.153 }, 00:25:30.153 "memory_domains": [ 00:25:30.153 { 00:25:30.153 "dma_device_id": "system", 00:25:30.153 "dma_device_type": 1 00:25:30.153 }, 00:25:30.153 { 00:25:30.153 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:30.153 "dma_device_type": 2 00:25:30.153 } 00:25:30.153 ], 00:25:30.153 "driver_specific": {} 00:25:30.153 } 00:25:30.153 ] 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.153 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:30.153 "name": "Existed_Raid", 00:25:30.153 "uuid": "da79bddb-7716-4558-acc0-00cd9de7521f", 00:25:30.153 "strip_size_kb": 0, 00:25:30.153 "state": "online", 00:25:30.153 "raid_level": "raid1", 00:25:30.153 "superblock": true, 00:25:30.153 "num_base_bdevs": 4, 00:25:30.153 "num_base_bdevs_discovered": 4, 00:25:30.153 "num_base_bdevs_operational": 4, 00:25:30.153 "base_bdevs_list": [ 00:25:30.153 { 00:25:30.153 "name": "NewBaseBdev", 00:25:30.153 "uuid": "1400b88a-2a60-4e39-99f1-c2effa87de04", 00:25:30.153 "is_configured": true, 00:25:30.153 "data_offset": 2048, 00:25:30.153 "data_size": 63488 00:25:30.154 }, 00:25:30.154 { 00:25:30.154 "name": "BaseBdev2", 00:25:30.154 "uuid": "150ef83d-f5f6-401b-88b5-e4bc90380427", 00:25:30.154 "is_configured": true, 00:25:30.154 "data_offset": 2048, 00:25:30.154 "data_size": 63488 00:25:30.154 }, 00:25:30.154 { 00:25:30.154 "name": "BaseBdev3", 00:25:30.154 "uuid": "d146c0d7-9bad-472b-9811-73868faa7ada", 00:25:30.154 "is_configured": true, 00:25:30.154 "data_offset": 2048, 00:25:30.154 "data_size": 63488 00:25:30.154 }, 00:25:30.154 { 00:25:30.154 "name": "BaseBdev4", 00:25:30.154 "uuid": "0473aaa1-4d55-4376-9527-77e6a5b1da3f", 00:25:30.154 "is_configured": true, 00:25:30.154 "data_offset": 2048, 00:25:30.154 "data_size": 63488 00:25:30.154 } 00:25:30.154 ] 00:25:30.154 }' 00:25:30.154 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:30.154 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:30.722 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:25:30.722 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:25:30.722 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:25:30.722 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:25:30.722 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:25:30.722 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:25:30.722 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:25:30.722 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.722 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:30.722 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:25:30.722 [2024-10-01 20:24:25.790326] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:25:30.722 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.722 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:30.722 "name": "Existed_Raid", 00:25:30.722 "aliases": [ 00:25:30.722 "da79bddb-7716-4558-acc0-00cd9de7521f" 00:25:30.722 ], 00:25:30.722 "product_name": "Raid Volume", 00:25:30.722 "block_size": 512, 00:25:30.722 "num_blocks": 63488, 00:25:30.722 "uuid": "da79bddb-7716-4558-acc0-00cd9de7521f", 00:25:30.722 "assigned_rate_limits": { 00:25:30.722 "rw_ios_per_sec": 0, 00:25:30.722 "rw_mbytes_per_sec": 0, 00:25:30.722 "r_mbytes_per_sec": 0, 00:25:30.722 "w_mbytes_per_sec": 0 00:25:30.722 }, 00:25:30.722 "claimed": false, 00:25:30.722 "zoned": false, 00:25:30.722 "supported_io_types": { 00:25:30.722 "read": true, 00:25:30.722 "write": true, 00:25:30.722 "unmap": false, 00:25:30.722 "flush": false, 00:25:30.722 "reset": true, 00:25:30.722 "nvme_admin": false, 00:25:30.722 "nvme_io": false, 00:25:30.722 "nvme_io_md": false, 00:25:30.722 "write_zeroes": true, 00:25:30.722 "zcopy": false, 00:25:30.722 "get_zone_info": false, 00:25:30.722 "zone_management": false, 00:25:30.722 "zone_append": false, 00:25:30.722 "compare": false, 00:25:30.722 "compare_and_write": false, 00:25:30.722 "abort": false, 00:25:30.722 "seek_hole": false, 00:25:30.722 "seek_data": false, 00:25:30.722 "copy": false, 00:25:30.722 "nvme_iov_md": false 00:25:30.722 }, 00:25:30.722 "memory_domains": [ 00:25:30.722 { 00:25:30.722 "dma_device_id": "system", 00:25:30.722 "dma_device_type": 1 00:25:30.722 }, 00:25:30.722 { 00:25:30.722 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:30.722 "dma_device_type": 2 00:25:30.722 }, 00:25:30.722 { 00:25:30.722 "dma_device_id": "system", 00:25:30.722 "dma_device_type": 1 00:25:30.722 }, 00:25:30.722 { 00:25:30.722 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:30.722 "dma_device_type": 2 00:25:30.722 }, 00:25:30.722 { 00:25:30.722 "dma_device_id": "system", 00:25:30.722 "dma_device_type": 1 00:25:30.722 }, 00:25:30.722 { 00:25:30.722 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:30.722 "dma_device_type": 2 00:25:30.722 }, 00:25:30.722 { 00:25:30.722 "dma_device_id": "system", 00:25:30.722 "dma_device_type": 1 00:25:30.722 }, 00:25:30.722 { 00:25:30.722 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:30.722 "dma_device_type": 2 00:25:30.722 } 00:25:30.722 ], 00:25:30.722 "driver_specific": { 00:25:30.722 "raid": { 00:25:30.722 "uuid": "da79bddb-7716-4558-acc0-00cd9de7521f", 00:25:30.722 "strip_size_kb": 0, 00:25:30.723 "state": "online", 00:25:30.723 "raid_level": "raid1", 00:25:30.723 "superblock": true, 00:25:30.723 "num_base_bdevs": 4, 00:25:30.723 "num_base_bdevs_discovered": 4, 00:25:30.723 "num_base_bdevs_operational": 4, 00:25:30.723 "base_bdevs_list": [ 00:25:30.723 { 00:25:30.723 "name": "NewBaseBdev", 00:25:30.723 "uuid": "1400b88a-2a60-4e39-99f1-c2effa87de04", 00:25:30.723 "is_configured": true, 00:25:30.723 "data_offset": 2048, 00:25:30.723 "data_size": 63488 00:25:30.723 }, 00:25:30.723 { 00:25:30.723 "name": "BaseBdev2", 00:25:30.723 "uuid": "150ef83d-f5f6-401b-88b5-e4bc90380427", 00:25:30.723 "is_configured": true, 00:25:30.723 "data_offset": 2048, 00:25:30.723 "data_size": 63488 00:25:30.723 }, 00:25:30.723 { 00:25:30.723 "name": "BaseBdev3", 00:25:30.723 "uuid": "d146c0d7-9bad-472b-9811-73868faa7ada", 00:25:30.723 "is_configured": true, 00:25:30.723 "data_offset": 2048, 00:25:30.723 "data_size": 63488 00:25:30.723 }, 00:25:30.723 { 00:25:30.723 "name": "BaseBdev4", 00:25:30.723 "uuid": "0473aaa1-4d55-4376-9527-77e6a5b1da3f", 00:25:30.723 "is_configured": true, 00:25:30.723 "data_offset": 2048, 00:25:30.723 "data_size": 63488 00:25:30.723 } 00:25:30.723 ] 00:25:30.723 } 00:25:30.723 } 00:25:30.723 }' 00:25:30.723 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:25:30.723 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:25:30.723 BaseBdev2 00:25:30.723 BaseBdev3 00:25:30.723 BaseBdev4' 00:25:30.723 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:30.723 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:25:30.723 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:30.723 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:25:30.723 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:30.723 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.723 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:30.723 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.981 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:30.981 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:30.981 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:30.981 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:25:30.981 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.981 20:24:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:30.981 20:24:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:30.981 [2024-10-01 20:24:26.150010] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:25:30.981 [2024-10-01 20:24:26.150049] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:25:30.981 [2024-10-01 20:24:26.150167] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:25:30.981 [2024-10-01 20:24:26.150578] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:25:30.981 [2024-10-01 20:24:26.150613] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 74720 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 74720 ']' 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 74720 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 74720 00:25:30.981 killing process with pid 74720 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 74720' 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 74720 00:25:30.981 [2024-10-01 20:24:26.192430] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:25:30.981 20:24:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 74720 00:25:31.602 [2024-10-01 20:24:26.549971] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:25:33.501 20:24:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:25:33.501 00:25:33.501 real 0m13.755s 00:25:33.501 user 0m22.089s 00:25:33.501 sys 0m1.973s 00:25:33.501 20:24:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:25:33.501 ************************************ 00:25:33.501 END TEST raid_state_function_test_sb 00:25:33.501 20:24:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:25:33.501 ************************************ 00:25:33.501 20:24:28 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 4 00:25:33.501 20:24:28 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:25:33.501 20:24:28 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:25:33.501 20:24:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:25:33.501 ************************************ 00:25:33.501 START TEST raid_superblock_test 00:25:33.501 ************************************ 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 4 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=75408 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 75408 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 75408 ']' 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:33.501 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:25:33.501 20:24:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:33.501 [2024-10-01 20:24:28.521384] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:25:33.501 [2024-10-01 20:24:28.521581] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid75408 ] 00:25:33.501 [2024-10-01 20:24:28.685545] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:33.760 [2024-10-01 20:24:28.930471] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:25:34.018 [2024-10-01 20:24:29.135293] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:34.019 [2024-10-01 20:24:29.135389] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:34.588 malloc1 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:34.588 [2024-10-01 20:24:29.588222] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:25:34.588 [2024-10-01 20:24:29.588316] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:34.588 [2024-10-01 20:24:29.588355] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:25:34.588 [2024-10-01 20:24:29.588378] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:34.588 [2024-10-01 20:24:29.591385] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:34.588 [2024-10-01 20:24:29.591435] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:25:34.588 pt1 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:34.588 malloc2 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:34.588 [2024-10-01 20:24:29.637153] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:25:34.588 [2024-10-01 20:24:29.637239] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:34.588 [2024-10-01 20:24:29.637285] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:25:34.588 [2024-10-01 20:24:29.637304] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:34.588 [2024-10-01 20:24:29.640257] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:34.588 [2024-10-01 20:24:29.640304] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:25:34.588 pt2 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:34.588 malloc3 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:34.588 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:34.588 [2024-10-01 20:24:29.686113] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:25:34.588 [2024-10-01 20:24:29.686188] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:34.588 [2024-10-01 20:24:29.686226] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:25:34.589 [2024-10-01 20:24:29.686245] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:34.589 [2024-10-01 20:24:29.689252] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:34.589 [2024-10-01 20:24:29.689297] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:25:34.589 pt3 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:34.589 malloc4 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:34.589 [2024-10-01 20:24:29.739107] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:25:34.589 [2024-10-01 20:24:29.739181] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:34.589 [2024-10-01 20:24:29.739214] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:25:34.589 [2024-10-01 20:24:29.739231] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:34.589 [2024-10-01 20:24:29.742094] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:34.589 [2024-10-01 20:24:29.742143] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:25:34.589 pt4 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:34.589 [2024-10-01 20:24:29.747152] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:25:34.589 [2024-10-01 20:24:29.749694] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:25:34.589 [2024-10-01 20:24:29.749819] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:25:34.589 [2024-10-01 20:24:29.749893] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:25:34.589 [2024-10-01 20:24:29.750173] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:25:34.589 [2024-10-01 20:24:29.750203] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:25:34.589 [2024-10-01 20:24:29.750562] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:25:34.589 [2024-10-01 20:24:29.750821] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:25:34.589 [2024-10-01 20:24:29.750866] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:25:34.589 [2024-10-01 20:24:29.751106] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:34.589 "name": "raid_bdev1", 00:25:34.589 "uuid": "e5054920-68a5-40ca-a5ad-baa7fec02133", 00:25:34.589 "strip_size_kb": 0, 00:25:34.589 "state": "online", 00:25:34.589 "raid_level": "raid1", 00:25:34.589 "superblock": true, 00:25:34.589 "num_base_bdevs": 4, 00:25:34.589 "num_base_bdevs_discovered": 4, 00:25:34.589 "num_base_bdevs_operational": 4, 00:25:34.589 "base_bdevs_list": [ 00:25:34.589 { 00:25:34.589 "name": "pt1", 00:25:34.589 "uuid": "00000000-0000-0000-0000-000000000001", 00:25:34.589 "is_configured": true, 00:25:34.589 "data_offset": 2048, 00:25:34.589 "data_size": 63488 00:25:34.589 }, 00:25:34.589 { 00:25:34.589 "name": "pt2", 00:25:34.589 "uuid": "00000000-0000-0000-0000-000000000002", 00:25:34.589 "is_configured": true, 00:25:34.589 "data_offset": 2048, 00:25:34.589 "data_size": 63488 00:25:34.589 }, 00:25:34.589 { 00:25:34.589 "name": "pt3", 00:25:34.589 "uuid": "00000000-0000-0000-0000-000000000003", 00:25:34.589 "is_configured": true, 00:25:34.589 "data_offset": 2048, 00:25:34.589 "data_size": 63488 00:25:34.589 }, 00:25:34.589 { 00:25:34.589 "name": "pt4", 00:25:34.589 "uuid": "00000000-0000-0000-0000-000000000004", 00:25:34.589 "is_configured": true, 00:25:34.589 "data_offset": 2048, 00:25:34.589 "data_size": 63488 00:25:34.589 } 00:25:34.589 ] 00:25:34.589 }' 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:34.589 20:24:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.157 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:25:35.157 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:25:35.157 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:25:35.157 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:25:35.157 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:25:35.157 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:25:35.157 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:25:35.157 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.157 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:25:35.157 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.157 [2024-10-01 20:24:30.211743] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:25:35.157 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:35.157 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:35.157 "name": "raid_bdev1", 00:25:35.157 "aliases": [ 00:25:35.157 "e5054920-68a5-40ca-a5ad-baa7fec02133" 00:25:35.157 ], 00:25:35.157 "product_name": "Raid Volume", 00:25:35.157 "block_size": 512, 00:25:35.157 "num_blocks": 63488, 00:25:35.157 "uuid": "e5054920-68a5-40ca-a5ad-baa7fec02133", 00:25:35.157 "assigned_rate_limits": { 00:25:35.157 "rw_ios_per_sec": 0, 00:25:35.157 "rw_mbytes_per_sec": 0, 00:25:35.157 "r_mbytes_per_sec": 0, 00:25:35.157 "w_mbytes_per_sec": 0 00:25:35.157 }, 00:25:35.157 "claimed": false, 00:25:35.157 "zoned": false, 00:25:35.157 "supported_io_types": { 00:25:35.157 "read": true, 00:25:35.157 "write": true, 00:25:35.157 "unmap": false, 00:25:35.157 "flush": false, 00:25:35.157 "reset": true, 00:25:35.157 "nvme_admin": false, 00:25:35.157 "nvme_io": false, 00:25:35.157 "nvme_io_md": false, 00:25:35.157 "write_zeroes": true, 00:25:35.157 "zcopy": false, 00:25:35.157 "get_zone_info": false, 00:25:35.157 "zone_management": false, 00:25:35.157 "zone_append": false, 00:25:35.157 "compare": false, 00:25:35.157 "compare_and_write": false, 00:25:35.157 "abort": false, 00:25:35.157 "seek_hole": false, 00:25:35.157 "seek_data": false, 00:25:35.157 "copy": false, 00:25:35.157 "nvme_iov_md": false 00:25:35.157 }, 00:25:35.157 "memory_domains": [ 00:25:35.157 { 00:25:35.157 "dma_device_id": "system", 00:25:35.157 "dma_device_type": 1 00:25:35.157 }, 00:25:35.157 { 00:25:35.157 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:35.157 "dma_device_type": 2 00:25:35.157 }, 00:25:35.157 { 00:25:35.157 "dma_device_id": "system", 00:25:35.157 "dma_device_type": 1 00:25:35.157 }, 00:25:35.157 { 00:25:35.157 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:35.157 "dma_device_type": 2 00:25:35.157 }, 00:25:35.157 { 00:25:35.157 "dma_device_id": "system", 00:25:35.157 "dma_device_type": 1 00:25:35.157 }, 00:25:35.157 { 00:25:35.157 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:35.157 "dma_device_type": 2 00:25:35.157 }, 00:25:35.157 { 00:25:35.157 "dma_device_id": "system", 00:25:35.157 "dma_device_type": 1 00:25:35.157 }, 00:25:35.157 { 00:25:35.157 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:35.157 "dma_device_type": 2 00:25:35.157 } 00:25:35.157 ], 00:25:35.157 "driver_specific": { 00:25:35.157 "raid": { 00:25:35.157 "uuid": "e5054920-68a5-40ca-a5ad-baa7fec02133", 00:25:35.157 "strip_size_kb": 0, 00:25:35.157 "state": "online", 00:25:35.157 "raid_level": "raid1", 00:25:35.157 "superblock": true, 00:25:35.157 "num_base_bdevs": 4, 00:25:35.157 "num_base_bdevs_discovered": 4, 00:25:35.157 "num_base_bdevs_operational": 4, 00:25:35.157 "base_bdevs_list": [ 00:25:35.157 { 00:25:35.157 "name": "pt1", 00:25:35.157 "uuid": "00000000-0000-0000-0000-000000000001", 00:25:35.157 "is_configured": true, 00:25:35.157 "data_offset": 2048, 00:25:35.157 "data_size": 63488 00:25:35.157 }, 00:25:35.157 { 00:25:35.157 "name": "pt2", 00:25:35.157 "uuid": "00000000-0000-0000-0000-000000000002", 00:25:35.157 "is_configured": true, 00:25:35.157 "data_offset": 2048, 00:25:35.158 "data_size": 63488 00:25:35.158 }, 00:25:35.158 { 00:25:35.158 "name": "pt3", 00:25:35.158 "uuid": "00000000-0000-0000-0000-000000000003", 00:25:35.158 "is_configured": true, 00:25:35.158 "data_offset": 2048, 00:25:35.158 "data_size": 63488 00:25:35.158 }, 00:25:35.158 { 00:25:35.158 "name": "pt4", 00:25:35.158 "uuid": "00000000-0000-0000-0000-000000000004", 00:25:35.158 "is_configured": true, 00:25:35.158 "data_offset": 2048, 00:25:35.158 "data_size": 63488 00:25:35.158 } 00:25:35.158 ] 00:25:35.158 } 00:25:35.158 } 00:25:35.158 }' 00:25:35.158 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:25:35.158 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:25:35.158 pt2 00:25:35.158 pt3 00:25:35.158 pt4' 00:25:35.158 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:35.158 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:25:35.158 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:35.158 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:25:35.158 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.158 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.158 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:35.158 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:35.158 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:35.158 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:35.158 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.417 [2024-10-01 20:24:30.559797] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=e5054920-68a5-40ca-a5ad-baa7fec02133 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z e5054920-68a5-40ca-a5ad-baa7fec02133 ']' 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.417 [2024-10-01 20:24:30.623404] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:25:35.417 [2024-10-01 20:24:30.623437] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:25:35.417 [2024-10-01 20:24:30.623540] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:25:35.417 [2024-10-01 20:24:30.623665] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:25:35.417 [2024-10-01 20:24:30.623693] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:25:35.417 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.677 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.677 [2024-10-01 20:24:30.779465] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:25:35.677 [2024-10-01 20:24:30.782086] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:25:35.677 [2024-10-01 20:24:30.782165] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:25:35.677 [2024-10-01 20:24:30.782223] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:25:35.677 [2024-10-01 20:24:30.782308] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:25:35.677 [2024-10-01 20:24:30.782381] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:25:35.678 [2024-10-01 20:24:30.782418] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:25:35.678 [2024-10-01 20:24:30.782455] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:25:35.678 [2024-10-01 20:24:30.782481] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:25:35.678 [2024-10-01 20:24:30.782499] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:25:35.678 request: 00:25:35.678 { 00:25:35.678 "name": "raid_bdev1", 00:25:35.678 "raid_level": "raid1", 00:25:35.678 "base_bdevs": [ 00:25:35.678 "malloc1", 00:25:35.678 "malloc2", 00:25:35.678 "malloc3", 00:25:35.678 "malloc4" 00:25:35.678 ], 00:25:35.678 "superblock": false, 00:25:35.678 "method": "bdev_raid_create", 00:25:35.678 "req_id": 1 00:25:35.678 } 00:25:35.678 Got JSON-RPC error response 00:25:35.678 response: 00:25:35.678 { 00:25:35.678 "code": -17, 00:25:35.678 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:25:35.678 } 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.678 [2024-10-01 20:24:30.843448] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:25:35.678 [2024-10-01 20:24:30.843527] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:35.678 [2024-10-01 20:24:30.843557] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:25:35.678 [2024-10-01 20:24:30.843577] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:35.678 [2024-10-01 20:24:30.846608] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:35.678 [2024-10-01 20:24:30.846695] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:25:35.678 [2024-10-01 20:24:30.846818] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:25:35.678 [2024-10-01 20:24:30.846896] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:25:35.678 pt1 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:35.678 "name": "raid_bdev1", 00:25:35.678 "uuid": "e5054920-68a5-40ca-a5ad-baa7fec02133", 00:25:35.678 "strip_size_kb": 0, 00:25:35.678 "state": "configuring", 00:25:35.678 "raid_level": "raid1", 00:25:35.678 "superblock": true, 00:25:35.678 "num_base_bdevs": 4, 00:25:35.678 "num_base_bdevs_discovered": 1, 00:25:35.678 "num_base_bdevs_operational": 4, 00:25:35.678 "base_bdevs_list": [ 00:25:35.678 { 00:25:35.678 "name": "pt1", 00:25:35.678 "uuid": "00000000-0000-0000-0000-000000000001", 00:25:35.678 "is_configured": true, 00:25:35.678 "data_offset": 2048, 00:25:35.678 "data_size": 63488 00:25:35.678 }, 00:25:35.678 { 00:25:35.678 "name": null, 00:25:35.678 "uuid": "00000000-0000-0000-0000-000000000002", 00:25:35.678 "is_configured": false, 00:25:35.678 "data_offset": 2048, 00:25:35.678 "data_size": 63488 00:25:35.678 }, 00:25:35.678 { 00:25:35.678 "name": null, 00:25:35.678 "uuid": "00000000-0000-0000-0000-000000000003", 00:25:35.678 "is_configured": false, 00:25:35.678 "data_offset": 2048, 00:25:35.678 "data_size": 63488 00:25:35.678 }, 00:25:35.678 { 00:25:35.678 "name": null, 00:25:35.678 "uuid": "00000000-0000-0000-0000-000000000004", 00:25:35.678 "is_configured": false, 00:25:35.678 "data_offset": 2048, 00:25:35.678 "data_size": 63488 00:25:35.678 } 00:25:35.678 ] 00:25:35.678 }' 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:35.678 20:24:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:36.247 [2024-10-01 20:24:31.331611] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:25:36.247 [2024-10-01 20:24:31.331706] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:36.247 [2024-10-01 20:24:31.331756] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:25:36.247 [2024-10-01 20:24:31.331779] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:36.247 [2024-10-01 20:24:31.332412] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:36.247 [2024-10-01 20:24:31.332457] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:25:36.247 [2024-10-01 20:24:31.332571] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:25:36.247 [2024-10-01 20:24:31.332623] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:25:36.247 pt2 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:36.247 [2024-10-01 20:24:31.339592] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:36.247 "name": "raid_bdev1", 00:25:36.247 "uuid": "e5054920-68a5-40ca-a5ad-baa7fec02133", 00:25:36.247 "strip_size_kb": 0, 00:25:36.247 "state": "configuring", 00:25:36.247 "raid_level": "raid1", 00:25:36.247 "superblock": true, 00:25:36.247 "num_base_bdevs": 4, 00:25:36.247 "num_base_bdevs_discovered": 1, 00:25:36.247 "num_base_bdevs_operational": 4, 00:25:36.247 "base_bdevs_list": [ 00:25:36.247 { 00:25:36.247 "name": "pt1", 00:25:36.247 "uuid": "00000000-0000-0000-0000-000000000001", 00:25:36.247 "is_configured": true, 00:25:36.247 "data_offset": 2048, 00:25:36.247 "data_size": 63488 00:25:36.247 }, 00:25:36.247 { 00:25:36.247 "name": null, 00:25:36.247 "uuid": "00000000-0000-0000-0000-000000000002", 00:25:36.247 "is_configured": false, 00:25:36.247 "data_offset": 0, 00:25:36.247 "data_size": 63488 00:25:36.247 }, 00:25:36.247 { 00:25:36.247 "name": null, 00:25:36.247 "uuid": "00000000-0000-0000-0000-000000000003", 00:25:36.247 "is_configured": false, 00:25:36.247 "data_offset": 2048, 00:25:36.247 "data_size": 63488 00:25:36.247 }, 00:25:36.247 { 00:25:36.247 "name": null, 00:25:36.247 "uuid": "00000000-0000-0000-0000-000000000004", 00:25:36.247 "is_configured": false, 00:25:36.247 "data_offset": 2048, 00:25:36.247 "data_size": 63488 00:25:36.247 } 00:25:36.247 ] 00:25:36.247 }' 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:36.247 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:36.891 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:25:36.891 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:25:36.891 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:25:36.891 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:36.891 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:36.891 [2024-10-01 20:24:31.843779] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:25:36.891 [2024-10-01 20:24:31.843867] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:36.891 [2024-10-01 20:24:31.843902] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:25:36.891 [2024-10-01 20:24:31.843920] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:36.891 [2024-10-01 20:24:31.844521] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:36.891 [2024-10-01 20:24:31.844560] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:25:36.891 [2024-10-01 20:24:31.844679] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:25:36.891 [2024-10-01 20:24:31.844731] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:25:36.891 pt2 00:25:36.891 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:36.891 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:25:36.891 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:25:36.891 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:25:36.891 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:36.891 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:36.891 [2024-10-01 20:24:31.851742] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:25:36.891 [2024-10-01 20:24:31.851814] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:36.891 [2024-10-01 20:24:31.851848] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:25:36.891 [2024-10-01 20:24:31.851864] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:36.891 [2024-10-01 20:24:31.852368] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:36.891 [2024-10-01 20:24:31.852400] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:25:36.891 [2024-10-01 20:24:31.852498] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:25:36.891 [2024-10-01 20:24:31.852538] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:25:36.891 pt3 00:25:36.891 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:36.891 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:25:36.891 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:36.892 [2024-10-01 20:24:31.859675] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:25:36.892 [2024-10-01 20:24:31.859890] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:36.892 [2024-10-01 20:24:31.859968] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:25:36.892 [2024-10-01 20:24:31.860118] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:36.892 [2024-10-01 20:24:31.860605] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:36.892 [2024-10-01 20:24:31.860643] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:25:36.892 [2024-10-01 20:24:31.860753] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:25:36.892 [2024-10-01 20:24:31.860784] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:25:36.892 [2024-10-01 20:24:31.860970] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:25:36.892 [2024-10-01 20:24:31.860988] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:25:36.892 [2024-10-01 20:24:31.861340] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:25:36.892 [2024-10-01 20:24:31.861684] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:25:36.892 [2024-10-01 20:24:31.861741] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:25:36.892 [2024-10-01 20:24:31.861928] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:36.892 pt4 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:36.892 "name": "raid_bdev1", 00:25:36.892 "uuid": "e5054920-68a5-40ca-a5ad-baa7fec02133", 00:25:36.892 "strip_size_kb": 0, 00:25:36.892 "state": "online", 00:25:36.892 "raid_level": "raid1", 00:25:36.892 "superblock": true, 00:25:36.892 "num_base_bdevs": 4, 00:25:36.892 "num_base_bdevs_discovered": 4, 00:25:36.892 "num_base_bdevs_operational": 4, 00:25:36.892 "base_bdevs_list": [ 00:25:36.892 { 00:25:36.892 "name": "pt1", 00:25:36.892 "uuid": "00000000-0000-0000-0000-000000000001", 00:25:36.892 "is_configured": true, 00:25:36.892 "data_offset": 2048, 00:25:36.892 "data_size": 63488 00:25:36.892 }, 00:25:36.892 { 00:25:36.892 "name": "pt2", 00:25:36.892 "uuid": "00000000-0000-0000-0000-000000000002", 00:25:36.892 "is_configured": true, 00:25:36.892 "data_offset": 2048, 00:25:36.892 "data_size": 63488 00:25:36.892 }, 00:25:36.892 { 00:25:36.892 "name": "pt3", 00:25:36.892 "uuid": "00000000-0000-0000-0000-000000000003", 00:25:36.892 "is_configured": true, 00:25:36.892 "data_offset": 2048, 00:25:36.892 "data_size": 63488 00:25:36.892 }, 00:25:36.892 { 00:25:36.892 "name": "pt4", 00:25:36.892 "uuid": "00000000-0000-0000-0000-000000000004", 00:25:36.892 "is_configured": true, 00:25:36.892 "data_offset": 2048, 00:25:36.892 "data_size": 63488 00:25:36.892 } 00:25:36.892 ] 00:25:36.892 }' 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:36.892 20:24:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:37.150 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:25:37.150 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:25:37.150 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:25:37.150 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:25:37.150 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:25:37.150 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:25:37.150 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:25:37.150 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:25:37.150 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:37.150 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:37.150 [2024-10-01 20:24:32.384333] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:25:37.409 "name": "raid_bdev1", 00:25:37.409 "aliases": [ 00:25:37.409 "e5054920-68a5-40ca-a5ad-baa7fec02133" 00:25:37.409 ], 00:25:37.409 "product_name": "Raid Volume", 00:25:37.409 "block_size": 512, 00:25:37.409 "num_blocks": 63488, 00:25:37.409 "uuid": "e5054920-68a5-40ca-a5ad-baa7fec02133", 00:25:37.409 "assigned_rate_limits": { 00:25:37.409 "rw_ios_per_sec": 0, 00:25:37.409 "rw_mbytes_per_sec": 0, 00:25:37.409 "r_mbytes_per_sec": 0, 00:25:37.409 "w_mbytes_per_sec": 0 00:25:37.409 }, 00:25:37.409 "claimed": false, 00:25:37.409 "zoned": false, 00:25:37.409 "supported_io_types": { 00:25:37.409 "read": true, 00:25:37.409 "write": true, 00:25:37.409 "unmap": false, 00:25:37.409 "flush": false, 00:25:37.409 "reset": true, 00:25:37.409 "nvme_admin": false, 00:25:37.409 "nvme_io": false, 00:25:37.409 "nvme_io_md": false, 00:25:37.409 "write_zeroes": true, 00:25:37.409 "zcopy": false, 00:25:37.409 "get_zone_info": false, 00:25:37.409 "zone_management": false, 00:25:37.409 "zone_append": false, 00:25:37.409 "compare": false, 00:25:37.409 "compare_and_write": false, 00:25:37.409 "abort": false, 00:25:37.409 "seek_hole": false, 00:25:37.409 "seek_data": false, 00:25:37.409 "copy": false, 00:25:37.409 "nvme_iov_md": false 00:25:37.409 }, 00:25:37.409 "memory_domains": [ 00:25:37.409 { 00:25:37.409 "dma_device_id": "system", 00:25:37.409 "dma_device_type": 1 00:25:37.409 }, 00:25:37.409 { 00:25:37.409 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:37.409 "dma_device_type": 2 00:25:37.409 }, 00:25:37.409 { 00:25:37.409 "dma_device_id": "system", 00:25:37.409 "dma_device_type": 1 00:25:37.409 }, 00:25:37.409 { 00:25:37.409 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:37.409 "dma_device_type": 2 00:25:37.409 }, 00:25:37.409 { 00:25:37.409 "dma_device_id": "system", 00:25:37.409 "dma_device_type": 1 00:25:37.409 }, 00:25:37.409 { 00:25:37.409 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:37.409 "dma_device_type": 2 00:25:37.409 }, 00:25:37.409 { 00:25:37.409 "dma_device_id": "system", 00:25:37.409 "dma_device_type": 1 00:25:37.409 }, 00:25:37.409 { 00:25:37.409 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:25:37.409 "dma_device_type": 2 00:25:37.409 } 00:25:37.409 ], 00:25:37.409 "driver_specific": { 00:25:37.409 "raid": { 00:25:37.409 "uuid": "e5054920-68a5-40ca-a5ad-baa7fec02133", 00:25:37.409 "strip_size_kb": 0, 00:25:37.409 "state": "online", 00:25:37.409 "raid_level": "raid1", 00:25:37.409 "superblock": true, 00:25:37.409 "num_base_bdevs": 4, 00:25:37.409 "num_base_bdevs_discovered": 4, 00:25:37.409 "num_base_bdevs_operational": 4, 00:25:37.409 "base_bdevs_list": [ 00:25:37.409 { 00:25:37.409 "name": "pt1", 00:25:37.409 "uuid": "00000000-0000-0000-0000-000000000001", 00:25:37.409 "is_configured": true, 00:25:37.409 "data_offset": 2048, 00:25:37.409 "data_size": 63488 00:25:37.409 }, 00:25:37.409 { 00:25:37.409 "name": "pt2", 00:25:37.409 "uuid": "00000000-0000-0000-0000-000000000002", 00:25:37.409 "is_configured": true, 00:25:37.409 "data_offset": 2048, 00:25:37.409 "data_size": 63488 00:25:37.409 }, 00:25:37.409 { 00:25:37.409 "name": "pt3", 00:25:37.409 "uuid": "00000000-0000-0000-0000-000000000003", 00:25:37.409 "is_configured": true, 00:25:37.409 "data_offset": 2048, 00:25:37.409 "data_size": 63488 00:25:37.409 }, 00:25:37.409 { 00:25:37.409 "name": "pt4", 00:25:37.409 "uuid": "00000000-0000-0000-0000-000000000004", 00:25:37.409 "is_configured": true, 00:25:37.409 "data_offset": 2048, 00:25:37.409 "data_size": 63488 00:25:37.409 } 00:25:37.409 ] 00:25:37.409 } 00:25:37.409 } 00:25:37.409 }' 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:25:37.409 pt2 00:25:37.409 pt3 00:25:37.409 pt4' 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:37.409 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:37.668 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:37.668 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:37.668 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:25:37.668 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:25:37.668 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:37.668 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:37.669 [2024-10-01 20:24:32.740371] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' e5054920-68a5-40ca-a5ad-baa7fec02133 '!=' e5054920-68a5-40ca-a5ad-baa7fec02133 ']' 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:37.669 [2024-10-01 20:24:32.784045] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:37.669 "name": "raid_bdev1", 00:25:37.669 "uuid": "e5054920-68a5-40ca-a5ad-baa7fec02133", 00:25:37.669 "strip_size_kb": 0, 00:25:37.669 "state": "online", 00:25:37.669 "raid_level": "raid1", 00:25:37.669 "superblock": true, 00:25:37.669 "num_base_bdevs": 4, 00:25:37.669 "num_base_bdevs_discovered": 3, 00:25:37.669 "num_base_bdevs_operational": 3, 00:25:37.669 "base_bdevs_list": [ 00:25:37.669 { 00:25:37.669 "name": null, 00:25:37.669 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:37.669 "is_configured": false, 00:25:37.669 "data_offset": 0, 00:25:37.669 "data_size": 63488 00:25:37.669 }, 00:25:37.669 { 00:25:37.669 "name": "pt2", 00:25:37.669 "uuid": "00000000-0000-0000-0000-000000000002", 00:25:37.669 "is_configured": true, 00:25:37.669 "data_offset": 2048, 00:25:37.669 "data_size": 63488 00:25:37.669 }, 00:25:37.669 { 00:25:37.669 "name": "pt3", 00:25:37.669 "uuid": "00000000-0000-0000-0000-000000000003", 00:25:37.669 "is_configured": true, 00:25:37.669 "data_offset": 2048, 00:25:37.669 "data_size": 63488 00:25:37.669 }, 00:25:37.669 { 00:25:37.669 "name": "pt4", 00:25:37.669 "uuid": "00000000-0000-0000-0000-000000000004", 00:25:37.669 "is_configured": true, 00:25:37.669 "data_offset": 2048, 00:25:37.669 "data_size": 63488 00:25:37.669 } 00:25:37.669 ] 00:25:37.669 }' 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:37.669 20:24:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:38.236 [2024-10-01 20:24:33.324121] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:25:38.236 [2024-10-01 20:24:33.324164] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:25:38.236 [2024-10-01 20:24:33.324274] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:25:38.236 [2024-10-01 20:24:33.324385] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:25:38.236 [2024-10-01 20:24:33.324404] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:38.236 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:38.236 [2024-10-01 20:24:33.416116] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:25:38.237 [2024-10-01 20:24:33.416321] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:38.237 [2024-10-01 20:24:33.416369] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:25:38.237 [2024-10-01 20:24:33.416389] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:38.237 [2024-10-01 20:24:33.419466] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:38.237 [2024-10-01 20:24:33.419628] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:25:38.237 [2024-10-01 20:24:33.419778] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:25:38.237 [2024-10-01 20:24:33.419845] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:25:38.237 pt2 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:38.237 "name": "raid_bdev1", 00:25:38.237 "uuid": "e5054920-68a5-40ca-a5ad-baa7fec02133", 00:25:38.237 "strip_size_kb": 0, 00:25:38.237 "state": "configuring", 00:25:38.237 "raid_level": "raid1", 00:25:38.237 "superblock": true, 00:25:38.237 "num_base_bdevs": 4, 00:25:38.237 "num_base_bdevs_discovered": 1, 00:25:38.237 "num_base_bdevs_operational": 3, 00:25:38.237 "base_bdevs_list": [ 00:25:38.237 { 00:25:38.237 "name": null, 00:25:38.237 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:38.237 "is_configured": false, 00:25:38.237 "data_offset": 2048, 00:25:38.237 "data_size": 63488 00:25:38.237 }, 00:25:38.237 { 00:25:38.237 "name": "pt2", 00:25:38.237 "uuid": "00000000-0000-0000-0000-000000000002", 00:25:38.237 "is_configured": true, 00:25:38.237 "data_offset": 2048, 00:25:38.237 "data_size": 63488 00:25:38.237 }, 00:25:38.237 { 00:25:38.237 "name": null, 00:25:38.237 "uuid": "00000000-0000-0000-0000-000000000003", 00:25:38.237 "is_configured": false, 00:25:38.237 "data_offset": 2048, 00:25:38.237 "data_size": 63488 00:25:38.237 }, 00:25:38.237 { 00:25:38.237 "name": null, 00:25:38.237 "uuid": "00000000-0000-0000-0000-000000000004", 00:25:38.237 "is_configured": false, 00:25:38.237 "data_offset": 2048, 00:25:38.237 "data_size": 63488 00:25:38.237 } 00:25:38.237 ] 00:25:38.237 }' 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:38.237 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:38.804 [2024-10-01 20:24:33.936304] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:25:38.804 [2024-10-01 20:24:33.936529] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:38.804 [2024-10-01 20:24:33.936735] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:25:38.804 [2024-10-01 20:24:33.936767] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:38.804 [2024-10-01 20:24:33.937425] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:38.804 [2024-10-01 20:24:33.937453] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:25:38.804 [2024-10-01 20:24:33.937581] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:25:38.804 [2024-10-01 20:24:33.937615] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:25:38.804 pt3 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:38.804 "name": "raid_bdev1", 00:25:38.804 "uuid": "e5054920-68a5-40ca-a5ad-baa7fec02133", 00:25:38.804 "strip_size_kb": 0, 00:25:38.804 "state": "configuring", 00:25:38.804 "raid_level": "raid1", 00:25:38.804 "superblock": true, 00:25:38.804 "num_base_bdevs": 4, 00:25:38.804 "num_base_bdevs_discovered": 2, 00:25:38.804 "num_base_bdevs_operational": 3, 00:25:38.804 "base_bdevs_list": [ 00:25:38.804 { 00:25:38.804 "name": null, 00:25:38.804 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:38.804 "is_configured": false, 00:25:38.804 "data_offset": 2048, 00:25:38.804 "data_size": 63488 00:25:38.804 }, 00:25:38.804 { 00:25:38.804 "name": "pt2", 00:25:38.804 "uuid": "00000000-0000-0000-0000-000000000002", 00:25:38.804 "is_configured": true, 00:25:38.804 "data_offset": 2048, 00:25:38.804 "data_size": 63488 00:25:38.804 }, 00:25:38.804 { 00:25:38.804 "name": "pt3", 00:25:38.804 "uuid": "00000000-0000-0000-0000-000000000003", 00:25:38.804 "is_configured": true, 00:25:38.804 "data_offset": 2048, 00:25:38.804 "data_size": 63488 00:25:38.804 }, 00:25:38.804 { 00:25:38.804 "name": null, 00:25:38.804 "uuid": "00000000-0000-0000-0000-000000000004", 00:25:38.804 "is_configured": false, 00:25:38.804 "data_offset": 2048, 00:25:38.804 "data_size": 63488 00:25:38.804 } 00:25:38.804 ] 00:25:38.804 }' 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:38.804 20:24:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:39.372 [2024-10-01 20:24:34.464439] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:25:39.372 [2024-10-01 20:24:34.464536] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:39.372 [2024-10-01 20:24:34.464577] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:25:39.372 [2024-10-01 20:24:34.464595] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:39.372 [2024-10-01 20:24:34.465259] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:39.372 [2024-10-01 20:24:34.465286] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:25:39.372 [2024-10-01 20:24:34.465400] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:25:39.372 [2024-10-01 20:24:34.465442] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:25:39.372 [2024-10-01 20:24:34.465630] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:25:39.372 [2024-10-01 20:24:34.465649] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:25:39.372 [2024-10-01 20:24:34.465991] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:25:39.372 [2024-10-01 20:24:34.466208] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:25:39.372 [2024-10-01 20:24:34.466231] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:25:39.372 [2024-10-01 20:24:34.466406] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:39.372 pt4 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:39.372 "name": "raid_bdev1", 00:25:39.372 "uuid": "e5054920-68a5-40ca-a5ad-baa7fec02133", 00:25:39.372 "strip_size_kb": 0, 00:25:39.372 "state": "online", 00:25:39.372 "raid_level": "raid1", 00:25:39.372 "superblock": true, 00:25:39.372 "num_base_bdevs": 4, 00:25:39.372 "num_base_bdevs_discovered": 3, 00:25:39.372 "num_base_bdevs_operational": 3, 00:25:39.372 "base_bdevs_list": [ 00:25:39.372 { 00:25:39.372 "name": null, 00:25:39.372 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:39.372 "is_configured": false, 00:25:39.372 "data_offset": 2048, 00:25:39.372 "data_size": 63488 00:25:39.372 }, 00:25:39.372 { 00:25:39.372 "name": "pt2", 00:25:39.372 "uuid": "00000000-0000-0000-0000-000000000002", 00:25:39.372 "is_configured": true, 00:25:39.372 "data_offset": 2048, 00:25:39.372 "data_size": 63488 00:25:39.372 }, 00:25:39.372 { 00:25:39.372 "name": "pt3", 00:25:39.372 "uuid": "00000000-0000-0000-0000-000000000003", 00:25:39.372 "is_configured": true, 00:25:39.372 "data_offset": 2048, 00:25:39.372 "data_size": 63488 00:25:39.372 }, 00:25:39.372 { 00:25:39.372 "name": "pt4", 00:25:39.372 "uuid": "00000000-0000-0000-0000-000000000004", 00:25:39.372 "is_configured": true, 00:25:39.372 "data_offset": 2048, 00:25:39.372 "data_size": 63488 00:25:39.372 } 00:25:39.372 ] 00:25:39.372 }' 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:39.372 20:24:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:39.940 20:24:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:25:39.940 20:24:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:39.940 20:24:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:39.940 [2024-10-01 20:24:34.996519] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:25:39.940 [2024-10-01 20:24:34.996560] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:25:39.940 [2024-10-01 20:24:34.996667] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:25:39.940 [2024-10-01 20:24:34.996788] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:25:39.940 [2024-10-01 20:24:34.996814] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:39.940 [2024-10-01 20:24:35.072531] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:25:39.940 [2024-10-01 20:24:35.072610] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:39.940 [2024-10-01 20:24:35.072640] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:25:39.940 [2024-10-01 20:24:35.072661] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:39.940 [2024-10-01 20:24:35.075794] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:39.940 [2024-10-01 20:24:35.075846] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:25:39.940 [2024-10-01 20:24:35.075963] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:25:39.940 [2024-10-01 20:24:35.076032] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:25:39.940 [2024-10-01 20:24:35.076200] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:25:39.940 [2024-10-01 20:24:35.076230] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:25:39.940 [2024-10-01 20:24:35.076258] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:25:39.940 [2024-10-01 20:24:35.076341] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:25:39.940 [2024-10-01 20:24:35.076493] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:25:39.940 pt1 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:39.940 "name": "raid_bdev1", 00:25:39.940 "uuid": "e5054920-68a5-40ca-a5ad-baa7fec02133", 00:25:39.940 "strip_size_kb": 0, 00:25:39.940 "state": "configuring", 00:25:39.940 "raid_level": "raid1", 00:25:39.940 "superblock": true, 00:25:39.940 "num_base_bdevs": 4, 00:25:39.940 "num_base_bdevs_discovered": 2, 00:25:39.940 "num_base_bdevs_operational": 3, 00:25:39.940 "base_bdevs_list": [ 00:25:39.940 { 00:25:39.940 "name": null, 00:25:39.940 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:39.940 "is_configured": false, 00:25:39.940 "data_offset": 2048, 00:25:39.940 "data_size": 63488 00:25:39.940 }, 00:25:39.940 { 00:25:39.940 "name": "pt2", 00:25:39.940 "uuid": "00000000-0000-0000-0000-000000000002", 00:25:39.940 "is_configured": true, 00:25:39.940 "data_offset": 2048, 00:25:39.940 "data_size": 63488 00:25:39.940 }, 00:25:39.940 { 00:25:39.940 "name": "pt3", 00:25:39.940 "uuid": "00000000-0000-0000-0000-000000000003", 00:25:39.940 "is_configured": true, 00:25:39.940 "data_offset": 2048, 00:25:39.940 "data_size": 63488 00:25:39.940 }, 00:25:39.940 { 00:25:39.940 "name": null, 00:25:39.940 "uuid": "00000000-0000-0000-0000-000000000004", 00:25:39.940 "is_configured": false, 00:25:39.940 "data_offset": 2048, 00:25:39.940 "data_size": 63488 00:25:39.940 } 00:25:39.940 ] 00:25:39.940 }' 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:39.940 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:40.507 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:25:40.507 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:40.507 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:40.507 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:25:40.507 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:40.507 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:25:40.507 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:25:40.507 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:40.507 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:40.507 [2024-10-01 20:24:35.664802] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:25:40.507 [2024-10-01 20:24:35.664893] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:40.507 [2024-10-01 20:24:35.664932] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:25:40.507 [2024-10-01 20:24:35.664951] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:40.507 [2024-10-01 20:24:35.665572] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:40.507 [2024-10-01 20:24:35.665600] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:25:40.507 [2024-10-01 20:24:35.665731] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:25:40.507 [2024-10-01 20:24:35.665767] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:25:40.507 [2024-10-01 20:24:35.665956] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:25:40.507 [2024-10-01 20:24:35.665974] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:25:40.507 [2024-10-01 20:24:35.666307] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:25:40.507 [2024-10-01 20:24:35.666497] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:25:40.507 [2024-10-01 20:24:35.666520] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:25:40.507 [2024-10-01 20:24:35.666705] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:40.507 pt4 00:25:40.507 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:40.508 "name": "raid_bdev1", 00:25:40.508 "uuid": "e5054920-68a5-40ca-a5ad-baa7fec02133", 00:25:40.508 "strip_size_kb": 0, 00:25:40.508 "state": "online", 00:25:40.508 "raid_level": "raid1", 00:25:40.508 "superblock": true, 00:25:40.508 "num_base_bdevs": 4, 00:25:40.508 "num_base_bdevs_discovered": 3, 00:25:40.508 "num_base_bdevs_operational": 3, 00:25:40.508 "base_bdevs_list": [ 00:25:40.508 { 00:25:40.508 "name": null, 00:25:40.508 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:40.508 "is_configured": false, 00:25:40.508 "data_offset": 2048, 00:25:40.508 "data_size": 63488 00:25:40.508 }, 00:25:40.508 { 00:25:40.508 "name": "pt2", 00:25:40.508 "uuid": "00000000-0000-0000-0000-000000000002", 00:25:40.508 "is_configured": true, 00:25:40.508 "data_offset": 2048, 00:25:40.508 "data_size": 63488 00:25:40.508 }, 00:25:40.508 { 00:25:40.508 "name": "pt3", 00:25:40.508 "uuid": "00000000-0000-0000-0000-000000000003", 00:25:40.508 "is_configured": true, 00:25:40.508 "data_offset": 2048, 00:25:40.508 "data_size": 63488 00:25:40.508 }, 00:25:40.508 { 00:25:40.508 "name": "pt4", 00:25:40.508 "uuid": "00000000-0000-0000-0000-000000000004", 00:25:40.508 "is_configured": true, 00:25:40.508 "data_offset": 2048, 00:25:40.508 "data_size": 63488 00:25:40.508 } 00:25:40.508 ] 00:25:40.508 }' 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:40.508 20:24:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:25:41.076 [2024-10-01 20:24:36.257295] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' e5054920-68a5-40ca-a5ad-baa7fec02133 '!=' e5054920-68a5-40ca-a5ad-baa7fec02133 ']' 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 75408 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 75408 ']' 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 75408 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:25:41.076 20:24:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 75408 00:25:41.335 20:24:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:25:41.335 killing process with pid 75408 00:25:41.335 20:24:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:25:41.335 20:24:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 75408' 00:25:41.335 20:24:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 75408 00:25:41.335 [2024-10-01 20:24:36.336280] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:25:41.335 20:24:36 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 75408 00:25:41.335 [2024-10-01 20:24:36.336407] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:25:41.335 [2024-10-01 20:24:36.336518] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:25:41.335 [2024-10-01 20:24:36.336542] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:25:41.595 [2024-10-01 20:24:36.694143] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:25:43.504 20:24:38 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:25:43.504 00:25:43.504 real 0m10.082s 00:25:43.504 user 0m16.001s 00:25:43.504 sys 0m1.444s 00:25:43.504 20:24:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:25:43.504 ************************************ 00:25:43.504 END TEST raid_superblock_test 00:25:43.504 ************************************ 00:25:43.504 20:24:38 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:25:43.504 20:24:38 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 4 read 00:25:43.504 20:24:38 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:25:43.504 20:24:38 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:25:43.504 20:24:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:25:43.504 ************************************ 00:25:43.504 START TEST raid_read_error_test 00:25:43.504 ************************************ 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 4 read 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.tmb5D1imD3 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=75911 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 75911 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 75911 ']' 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:43.504 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:25:43.504 20:24:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:43.504 [2024-10-01 20:24:38.689971] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:25:43.504 [2024-10-01 20:24:38.690178] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid75911 ] 00:25:43.763 [2024-10-01 20:24:38.861716] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:44.021 [2024-10-01 20:24:39.109269] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:25:44.279 [2024-10-01 20:24:39.324118] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:44.279 [2024-10-01 20:24:39.324234] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:44.538 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:25:44.538 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:25:44.538 20:24:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:25:44.538 20:24:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:25:44.538 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:44.538 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:44.798 BaseBdev1_malloc 00:25:44.798 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:44.798 20:24:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:25:44.798 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:44.798 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:44.798 true 00:25:44.798 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:44.798 20:24:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:25:44.798 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:44.798 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:44.798 [2024-10-01 20:24:39.815242] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:25:44.798 [2024-10-01 20:24:39.815321] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:44.799 [2024-10-01 20:24:39.815350] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:25:44.799 [2024-10-01 20:24:39.815370] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:44.799 [2024-10-01 20:24:39.818329] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:44.799 BaseBdev1 00:25:44.799 [2024-10-01 20:24:39.818533] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:44.799 BaseBdev2_malloc 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:44.799 true 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:44.799 [2024-10-01 20:24:39.878290] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:25:44.799 [2024-10-01 20:24:39.878365] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:44.799 [2024-10-01 20:24:39.878394] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:25:44.799 [2024-10-01 20:24:39.878413] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:44.799 [2024-10-01 20:24:39.881489] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:44.799 BaseBdev2 00:25:44.799 [2024-10-01 20:24:39.881762] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:44.799 BaseBdev3_malloc 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:44.799 true 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:44.799 [2024-10-01 20:24:39.941508] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:25:44.799 [2024-10-01 20:24:39.941582] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:44.799 [2024-10-01 20:24:39.941609] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:25:44.799 [2024-10-01 20:24:39.941628] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:44.799 [2024-10-01 20:24:39.944537] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:44.799 [2024-10-01 20:24:39.944591] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:25:44.799 BaseBdev3 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:44.799 BaseBdev4_malloc 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:44.799 true 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:44.799 20:24:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:44.799 [2024-10-01 20:24:40.002273] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:25:44.799 [2024-10-01 20:24:40.002345] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:44.799 [2024-10-01 20:24:40.002377] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:25:44.799 [2024-10-01 20:24:40.002399] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:44.799 [2024-10-01 20:24:40.005399] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:44.799 [2024-10-01 20:24:40.005458] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:25:44.799 BaseBdev4 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:44.799 [2024-10-01 20:24:40.010426] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:25:44.799 [2024-10-01 20:24:40.013025] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:25:44.799 [2024-10-01 20:24:40.013138] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:25:44.799 [2024-10-01 20:24:40.013255] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:25:44.799 [2024-10-01 20:24:40.013602] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:25:44.799 [2024-10-01 20:24:40.013636] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:25:44.799 [2024-10-01 20:24:40.013969] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:25:44.799 [2024-10-01 20:24:40.014207] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:25:44.799 [2024-10-01 20:24:40.014233] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:25:44.799 [2024-10-01 20:24:40.014483] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:44.799 20:24:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:45.059 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:45.059 "name": "raid_bdev1", 00:25:45.059 "uuid": "d04ec133-9851-4eb4-b339-2bff3bc9af91", 00:25:45.059 "strip_size_kb": 0, 00:25:45.059 "state": "online", 00:25:45.059 "raid_level": "raid1", 00:25:45.059 "superblock": true, 00:25:45.059 "num_base_bdevs": 4, 00:25:45.059 "num_base_bdevs_discovered": 4, 00:25:45.059 "num_base_bdevs_operational": 4, 00:25:45.059 "base_bdevs_list": [ 00:25:45.059 { 00:25:45.059 "name": "BaseBdev1", 00:25:45.059 "uuid": "32606064-82fb-5ef4-a30f-4760d8d33d7f", 00:25:45.059 "is_configured": true, 00:25:45.059 "data_offset": 2048, 00:25:45.059 "data_size": 63488 00:25:45.059 }, 00:25:45.059 { 00:25:45.059 "name": "BaseBdev2", 00:25:45.059 "uuid": "9d356f64-87a0-54de-a655-2d32d942ee6f", 00:25:45.059 "is_configured": true, 00:25:45.059 "data_offset": 2048, 00:25:45.059 "data_size": 63488 00:25:45.059 }, 00:25:45.059 { 00:25:45.059 "name": "BaseBdev3", 00:25:45.059 "uuid": "c2394503-a1f2-59b1-9ab5-cc3f8e866a86", 00:25:45.059 "is_configured": true, 00:25:45.059 "data_offset": 2048, 00:25:45.059 "data_size": 63488 00:25:45.059 }, 00:25:45.059 { 00:25:45.059 "name": "BaseBdev4", 00:25:45.059 "uuid": "533a6f5c-91c2-555e-9678-1e95f90f2627", 00:25:45.059 "is_configured": true, 00:25:45.059 "data_offset": 2048, 00:25:45.059 "data_size": 63488 00:25:45.059 } 00:25:45.059 ] 00:25:45.059 }' 00:25:45.059 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:45.059 20:24:40 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:45.318 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:25:45.318 20:24:40 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:25:45.576 [2024-10-01 20:24:40.672051] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:46.513 "name": "raid_bdev1", 00:25:46.513 "uuid": "d04ec133-9851-4eb4-b339-2bff3bc9af91", 00:25:46.513 "strip_size_kb": 0, 00:25:46.513 "state": "online", 00:25:46.513 "raid_level": "raid1", 00:25:46.513 "superblock": true, 00:25:46.513 "num_base_bdevs": 4, 00:25:46.513 "num_base_bdevs_discovered": 4, 00:25:46.513 "num_base_bdevs_operational": 4, 00:25:46.513 "base_bdevs_list": [ 00:25:46.513 { 00:25:46.513 "name": "BaseBdev1", 00:25:46.513 "uuid": "32606064-82fb-5ef4-a30f-4760d8d33d7f", 00:25:46.513 "is_configured": true, 00:25:46.513 "data_offset": 2048, 00:25:46.513 "data_size": 63488 00:25:46.513 }, 00:25:46.513 { 00:25:46.513 "name": "BaseBdev2", 00:25:46.513 "uuid": "9d356f64-87a0-54de-a655-2d32d942ee6f", 00:25:46.513 "is_configured": true, 00:25:46.513 "data_offset": 2048, 00:25:46.513 "data_size": 63488 00:25:46.513 }, 00:25:46.513 { 00:25:46.513 "name": "BaseBdev3", 00:25:46.513 "uuid": "c2394503-a1f2-59b1-9ab5-cc3f8e866a86", 00:25:46.513 "is_configured": true, 00:25:46.513 "data_offset": 2048, 00:25:46.513 "data_size": 63488 00:25:46.513 }, 00:25:46.513 { 00:25:46.513 "name": "BaseBdev4", 00:25:46.513 "uuid": "533a6f5c-91c2-555e-9678-1e95f90f2627", 00:25:46.513 "is_configured": true, 00:25:46.513 "data_offset": 2048, 00:25:46.513 "data_size": 63488 00:25:46.513 } 00:25:46.513 ] 00:25:46.513 }' 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:46.513 20:24:41 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:47.079 20:24:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:25:47.079 20:24:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:47.080 20:24:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:47.080 [2024-10-01 20:24:42.065431] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:25:47.080 [2024-10-01 20:24:42.065477] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:25:47.080 [2024-10-01 20:24:42.069076] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:25:47.080 [2024-10-01 20:24:42.069197] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:47.080 [2024-10-01 20:24:42.069472] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:25:47.080 [2024-10-01 20:24:42.069509] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:25:47.080 { 00:25:47.080 "results": [ 00:25:47.080 { 00:25:47.080 "job": "raid_bdev1", 00:25:47.080 "core_mask": "0x1", 00:25:47.080 "workload": "randrw", 00:25:47.080 "percentage": 50, 00:25:47.080 "status": "finished", 00:25:47.080 "queue_depth": 1, 00:25:47.080 "io_size": 131072, 00:25:47.080 "runtime": 1.390712, 00:25:47.080 "iops": 6954.710968194709, 00:25:47.080 "mibps": 869.3388710243386, 00:25:47.080 "io_failed": 0, 00:25:47.080 "io_timeout": 0, 00:25:47.080 "avg_latency_us": 138.97705391382812, 00:25:47.080 "min_latency_us": 43.52, 00:25:47.080 "max_latency_us": 1936.290909090909 00:25:47.080 } 00:25:47.080 ], 00:25:47.080 "core_count": 1 00:25:47.080 } 00:25:47.080 20:24:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:47.080 20:24:42 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 75911 00:25:47.080 20:24:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 75911 ']' 00:25:47.080 20:24:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 75911 00:25:47.080 20:24:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:25:47.080 20:24:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:25:47.080 20:24:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 75911 00:25:47.080 killing process with pid 75911 00:25:47.080 20:24:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:25:47.080 20:24:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:25:47.080 20:24:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 75911' 00:25:47.080 20:24:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 75911 00:25:47.080 [2024-10-01 20:24:42.104619] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:25:47.080 20:24:42 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 75911 00:25:47.382 [2024-10-01 20:24:42.414181] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:25:49.282 20:24:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.tmb5D1imD3 00:25:49.282 20:24:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:25:49.282 20:24:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:25:49.282 20:24:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:25:49.282 20:24:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:25:49.282 20:24:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:25:49.282 20:24:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:25:49.282 20:24:44 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:25:49.282 00:25:49.282 real 0m5.749s 00:25:49.282 user 0m6.768s 00:25:49.282 sys 0m0.702s 00:25:49.282 ************************************ 00:25:49.282 END TEST raid_read_error_test 00:25:49.282 ************************************ 00:25:49.282 20:24:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:25:49.282 20:24:44 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:49.282 20:24:44 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 4 write 00:25:49.282 20:24:44 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:25:49.282 20:24:44 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:25:49.282 20:24:44 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:25:49.282 ************************************ 00:25:49.282 START TEST raid_write_error_test 00:25:49.282 ************************************ 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 4 write 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.fZTld7bqNp 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76069 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76069 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 76069 ']' 00:25:49.282 20:24:44 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:25:49.283 20:24:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:49.283 20:24:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:25:49.283 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:49.283 20:24:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:49.283 20:24:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:25:49.283 20:24:44 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:49.283 [2024-10-01 20:24:44.492502] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:25:49.283 [2024-10-01 20:24:44.492701] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76069 ] 00:25:49.542 [2024-10-01 20:24:44.669171] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:49.801 [2024-10-01 20:24:44.900345] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:25:50.060 [2024-10-01 20:24:45.084758] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:50.060 [2024-10-01 20:24:45.085215] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:50.319 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:25:50.319 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:25:50.319 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:25:50.319 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:25:50.319 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:50.319 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:50.319 BaseBdev1_malloc 00:25:50.319 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:50.319 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:25:50.320 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:50.320 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:50.320 true 00:25:50.320 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:50.320 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:25:50.320 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:50.320 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:50.320 [2024-10-01 20:24:45.554296] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:25:50.320 [2024-10-01 20:24:45.554369] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:50.320 [2024-10-01 20:24:45.554416] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:25:50.320 [2024-10-01 20:24:45.554436] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:50.320 [2024-10-01 20:24:45.557485] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:50.320 [2024-10-01 20:24:45.557543] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:25:50.320 BaseBdev1 00:25:50.320 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:50.320 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:25:50.320 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:25:50.320 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:50.320 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:50.579 BaseBdev2_malloc 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:50.579 true 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:50.579 [2024-10-01 20:24:45.616702] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:25:50.579 [2024-10-01 20:24:45.616814] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:50.579 [2024-10-01 20:24:45.616844] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:25:50.579 [2024-10-01 20:24:45.616880] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:50.579 [2024-10-01 20:24:45.619977] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:50.579 [2024-10-01 20:24:45.620030] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:25:50.579 BaseBdev2 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:50.579 BaseBdev3_malloc 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:50.579 true 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:50.579 [2024-10-01 20:24:45.671046] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:25:50.579 [2024-10-01 20:24:45.671312] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:50.579 [2024-10-01 20:24:45.671352] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:25:50.579 [2024-10-01 20:24:45.671372] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:50.579 [2024-10-01 20:24:45.674444] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:50.579 [2024-10-01 20:24:45.674660] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:25:50.579 BaseBdev3 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:50.579 BaseBdev4_malloc 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:50.579 true 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:50.579 [2024-10-01 20:24:45.728875] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:25:50.579 [2024-10-01 20:24:45.729146] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:50.579 [2024-10-01 20:24:45.729218] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:25:50.579 [2024-10-01 20:24:45.729274] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:50.579 [2024-10-01 20:24:45.732376] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:50.579 BaseBdev4 00:25:50.579 [2024-10-01 20:24:45.732594] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:50.579 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:50.580 [2024-10-01 20:24:45.736974] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:25:50.580 [2024-10-01 20:24:45.739700] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:25:50.580 [2024-10-01 20:24:45.739836] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:25:50.580 [2024-10-01 20:24:45.739935] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:25:50.580 [2024-10-01 20:24:45.740256] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008580 00:25:50.580 [2024-10-01 20:24:45.740279] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:25:50.580 [2024-10-01 20:24:45.740568] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:25:50.580 [2024-10-01 20:24:45.740801] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008580 00:25:50.580 [2024-10-01 20:24:45.740818] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008580 00:25:50.580 [2024-10-01 20:24:45.741044] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:50.580 "name": "raid_bdev1", 00:25:50.580 "uuid": "630332a1-ad67-4080-bd47-f8aec1f466bb", 00:25:50.580 "strip_size_kb": 0, 00:25:50.580 "state": "online", 00:25:50.580 "raid_level": "raid1", 00:25:50.580 "superblock": true, 00:25:50.580 "num_base_bdevs": 4, 00:25:50.580 "num_base_bdevs_discovered": 4, 00:25:50.580 "num_base_bdevs_operational": 4, 00:25:50.580 "base_bdevs_list": [ 00:25:50.580 { 00:25:50.580 "name": "BaseBdev1", 00:25:50.580 "uuid": "d7ad36fb-c01a-57d3-835e-fbdd340d63c8", 00:25:50.580 "is_configured": true, 00:25:50.580 "data_offset": 2048, 00:25:50.580 "data_size": 63488 00:25:50.580 }, 00:25:50.580 { 00:25:50.580 "name": "BaseBdev2", 00:25:50.580 "uuid": "5186eaef-2140-5de0-8e9b-10399a2cee71", 00:25:50.580 "is_configured": true, 00:25:50.580 "data_offset": 2048, 00:25:50.580 "data_size": 63488 00:25:50.580 }, 00:25:50.580 { 00:25:50.580 "name": "BaseBdev3", 00:25:50.580 "uuid": "cf6976aa-0b96-56ab-b349-a60f48323f70", 00:25:50.580 "is_configured": true, 00:25:50.580 "data_offset": 2048, 00:25:50.580 "data_size": 63488 00:25:50.580 }, 00:25:50.580 { 00:25:50.580 "name": "BaseBdev4", 00:25:50.580 "uuid": "5fd8c7e0-7d6b-5dea-a644-ae48f38f81ca", 00:25:50.580 "is_configured": true, 00:25:50.580 "data_offset": 2048, 00:25:50.580 "data_size": 63488 00:25:50.580 } 00:25:50.580 ] 00:25:50.580 }' 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:50.580 20:24:45 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:51.146 20:24:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:25:51.146 20:24:46 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:25:51.146 [2024-10-01 20:24:46.378865] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:52.084 [2024-10-01 20:24:47.257273] bdev_raid.c:2272:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:25:52.084 [2024-10-01 20:24:47.257496] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:25:52.084 [2024-10-01 20:24:47.257837] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006080 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:52.084 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:52.085 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:52.085 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:52.085 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:52.085 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:52.085 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:52.085 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:52.085 "name": "raid_bdev1", 00:25:52.085 "uuid": "630332a1-ad67-4080-bd47-f8aec1f466bb", 00:25:52.085 "strip_size_kb": 0, 00:25:52.085 "state": "online", 00:25:52.085 "raid_level": "raid1", 00:25:52.085 "superblock": true, 00:25:52.085 "num_base_bdevs": 4, 00:25:52.085 "num_base_bdevs_discovered": 3, 00:25:52.085 "num_base_bdevs_operational": 3, 00:25:52.085 "base_bdevs_list": [ 00:25:52.085 { 00:25:52.085 "name": null, 00:25:52.085 "uuid": "00000000-0000-0000-0000-000000000000", 00:25:52.085 "is_configured": false, 00:25:52.085 "data_offset": 0, 00:25:52.085 "data_size": 63488 00:25:52.085 }, 00:25:52.085 { 00:25:52.085 "name": "BaseBdev2", 00:25:52.085 "uuid": "5186eaef-2140-5de0-8e9b-10399a2cee71", 00:25:52.085 "is_configured": true, 00:25:52.085 "data_offset": 2048, 00:25:52.085 "data_size": 63488 00:25:52.085 }, 00:25:52.085 { 00:25:52.085 "name": "BaseBdev3", 00:25:52.085 "uuid": "cf6976aa-0b96-56ab-b349-a60f48323f70", 00:25:52.085 "is_configured": true, 00:25:52.085 "data_offset": 2048, 00:25:52.085 "data_size": 63488 00:25:52.085 }, 00:25:52.085 { 00:25:52.085 "name": "BaseBdev4", 00:25:52.085 "uuid": "5fd8c7e0-7d6b-5dea-a644-ae48f38f81ca", 00:25:52.085 "is_configured": true, 00:25:52.085 "data_offset": 2048, 00:25:52.085 "data_size": 63488 00:25:52.085 } 00:25:52.085 ] 00:25:52.085 }' 00:25:52.085 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:52.085 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:52.651 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:25:52.651 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:52.651 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:52.651 [2024-10-01 20:24:47.774745] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:25:52.651 [2024-10-01 20:24:47.774785] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:25:52.651 [2024-10-01 20:24:47.778153] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:25:52.651 { 00:25:52.651 "results": [ 00:25:52.651 { 00:25:52.651 "job": "raid_bdev1", 00:25:52.651 "core_mask": "0x1", 00:25:52.651 "workload": "randrw", 00:25:52.651 "percentage": 50, 00:25:52.651 "status": "finished", 00:25:52.651 "queue_depth": 1, 00:25:52.651 "io_size": 131072, 00:25:52.651 "runtime": 1.393089, 00:25:52.651 "iops": 7690.822337984149, 00:25:52.651 "mibps": 961.3527922480187, 00:25:52.651 "io_failed": 0, 00:25:52.651 "io_timeout": 0, 00:25:52.651 "avg_latency_us": 125.50800006788062, 00:25:52.651 "min_latency_us": 41.192727272727275, 00:25:52.651 "max_latency_us": 2100.130909090909 00:25:52.651 } 00:25:52.651 ], 00:25:52.651 "core_count": 1 00:25:52.651 } 00:25:52.651 [2024-10-01 20:24:47.778382] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:52.651 [2024-10-01 20:24:47.778548] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:25:52.651 [2024-10-01 20:24:47.778574] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state offline 00:25:52.651 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:52.651 20:24:47 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76069 00:25:52.651 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 76069 ']' 00:25:52.651 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 76069 00:25:52.651 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:25:52.651 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:25:52.651 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 76069 00:25:52.651 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:25:52.651 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:25:52.651 killing process with pid 76069 00:25:52.651 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 76069' 00:25:52.651 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 76069 00:25:52.651 [2024-10-01 20:24:47.816847] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:25:52.651 20:24:47 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 76069 00:25:52.910 [2024-10-01 20:24:48.113989] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:25:54.813 20:24:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.fZTld7bqNp 00:25:54.813 20:24:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:25:54.813 20:24:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:25:54.813 20:24:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:25:54.813 20:24:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:25:54.813 20:24:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:25:54.813 20:24:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:25:54.813 20:24:50 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:25:54.813 00:25:54.813 real 0m5.642s 00:25:54.813 user 0m6.604s 00:25:54.813 sys 0m0.720s 00:25:54.813 ************************************ 00:25:54.813 END TEST raid_write_error_test 00:25:54.813 ************************************ 00:25:54.813 20:24:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:25:54.813 20:24:50 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:25:54.813 20:24:50 bdev_raid -- bdev/bdev_raid.sh@976 -- # '[' true = true ']' 00:25:54.813 20:24:50 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:25:54.813 20:24:50 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 2 false false true 00:25:54.813 20:24:50 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:25:54.813 20:24:50 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:25:54.813 20:24:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:25:54.813 ************************************ 00:25:54.813 START TEST raid_rebuild_test 00:25:54.813 ************************************ 00:25:54.813 20:24:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 false false true 00:25:54.813 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:25:54.813 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:25:54.813 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:25:54.813 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:25:54.813 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:25:55.071 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=76217 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 76217 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@831 -- # '[' -z 76217 ']' 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:55.071 20:24:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:25:55.072 20:24:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:55.072 20:24:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:25:55.072 20:24:50 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:25:55.072 [2024-10-01 20:24:50.184965] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:25:55.072 [2024-10-01 20:24:50.185414] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --matchI/O size of 3145728 is greater than zero copy threshold (65536). 00:25:55.072 Zero copy mechanism will not be used. 00:25:55.072 -allocations --file-prefix=spdk_pid76217 ] 00:25:55.330 [2024-10-01 20:24:50.364669] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:55.652 [2024-10-01 20:24:50.651817] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:25:55.652 [2024-10-01 20:24:50.858544] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:55.653 [2024-10-01 20:24:50.858864] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # return 0 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:25:56.220 BaseBdev1_malloc 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:25:56.220 [2024-10-01 20:24:51.327174] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:25:56.220 [2024-10-01 20:24:51.327436] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:56.220 [2024-10-01 20:24:51.327595] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:25:56.220 [2024-10-01 20:24:51.327777] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:56.220 [2024-10-01 20:24:51.330961] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:56.220 BaseBdev1 00:25:56.220 [2024-10-01 20:24:51.331139] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:25:56.220 BaseBdev2_malloc 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:25:56.220 [2024-10-01 20:24:51.384894] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:25:56.220 [2024-10-01 20:24:51.385001] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:56.220 [2024-10-01 20:24:51.385043] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:25:56.220 [2024-10-01 20:24:51.385063] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:56.220 [2024-10-01 20:24:51.388045] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:56.220 [2024-10-01 20:24:51.388218] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:25:56.220 BaseBdev2 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:25:56.220 spare_malloc 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:25:56.220 spare_delay 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:56.220 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:25:56.220 [2024-10-01 20:24:51.446304] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:25:56.220 [2024-10-01 20:24:51.446396] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:25:56.220 [2024-10-01 20:24:51.446445] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:25:56.220 [2024-10-01 20:24:51.446466] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:25:56.221 [2024-10-01 20:24:51.449499] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:25:56.221 [2024-10-01 20:24:51.449554] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:25:56.221 spare 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:25:56.221 [2024-10-01 20:24:51.454470] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:25:56.221 [2024-10-01 20:24:51.457227] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:25:56.221 [2024-10-01 20:24:51.457384] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:25:56.221 [2024-10-01 20:24:51.457406] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:25:56.221 [2024-10-01 20:24:51.457932] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:25:56.221 [2024-10-01 20:24:51.458293] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:25:56.221 [2024-10-01 20:24:51.458435] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:25:56.221 [2024-10-01 20:24:51.458724] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:25:56.221 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:25:56.479 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:56.479 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:25:56.479 "name": "raid_bdev1", 00:25:56.479 "uuid": "e463d538-8861-4ac1-82e4-8d298a9ec734", 00:25:56.479 "strip_size_kb": 0, 00:25:56.479 "state": "online", 00:25:56.479 "raid_level": "raid1", 00:25:56.479 "superblock": false, 00:25:56.479 "num_base_bdevs": 2, 00:25:56.479 "num_base_bdevs_discovered": 2, 00:25:56.479 "num_base_bdevs_operational": 2, 00:25:56.479 "base_bdevs_list": [ 00:25:56.479 { 00:25:56.479 "name": "BaseBdev1", 00:25:56.479 "uuid": "15dac736-893f-5d68-8553-b787dd70ab71", 00:25:56.479 "is_configured": true, 00:25:56.479 "data_offset": 0, 00:25:56.479 "data_size": 65536 00:25:56.479 }, 00:25:56.479 { 00:25:56.479 "name": "BaseBdev2", 00:25:56.479 "uuid": "4e9a1435-f39b-5e3f-8c2c-840c0d05770b", 00:25:56.479 "is_configured": true, 00:25:56.479 "data_offset": 0, 00:25:56.479 "data_size": 65536 00:25:56.479 } 00:25:56.479 ] 00:25:56.479 }' 00:25:56.479 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:25:56.479 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:25:56.737 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:25:56.737 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:56.737 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:25:56.737 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:25:56.737 [2024-10-01 20:24:51.951238] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:25:56.737 20:24:51 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:56.996 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:25:56.996 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:25:56.996 20:24:51 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:25:56.996 20:24:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:25:57.255 [2024-10-01 20:24:52.391080] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:25:57.255 /dev/nbd0 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:25:57.255 1+0 records in 00:25:57.255 1+0 records out 00:25:57.255 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000481483 s, 8.5 MB/s 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:25:57.255 20:24:52 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:26:05.444 65536+0 records in 00:26:05.444 65536+0 records out 00:26:05.444 33554432 bytes (34 MB, 32 MiB) copied, 6.84954 s, 4.9 MB/s 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:26:05.444 [2024-10-01 20:24:59.624505] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:05.444 [2024-10-01 20:24:59.660627] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:05.444 "name": "raid_bdev1", 00:26:05.444 "uuid": "e463d538-8861-4ac1-82e4-8d298a9ec734", 00:26:05.444 "strip_size_kb": 0, 00:26:05.444 "state": "online", 00:26:05.444 "raid_level": "raid1", 00:26:05.444 "superblock": false, 00:26:05.444 "num_base_bdevs": 2, 00:26:05.444 "num_base_bdevs_discovered": 1, 00:26:05.444 "num_base_bdevs_operational": 1, 00:26:05.444 "base_bdevs_list": [ 00:26:05.444 { 00:26:05.444 "name": null, 00:26:05.444 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:05.444 "is_configured": false, 00:26:05.444 "data_offset": 0, 00:26:05.444 "data_size": 65536 00:26:05.444 }, 00:26:05.444 { 00:26:05.444 "name": "BaseBdev2", 00:26:05.444 "uuid": "4e9a1435-f39b-5e3f-8c2c-840c0d05770b", 00:26:05.444 "is_configured": true, 00:26:05.444 "data_offset": 0, 00:26:05.444 "data_size": 65536 00:26:05.444 } 00:26:05.444 ] 00:26:05.444 }' 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:05.444 20:24:59 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:05.444 20:25:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:26:05.444 20:25:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:05.444 20:25:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:05.444 [2024-10-01 20:25:00.200835] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:26:05.444 [2024-10-01 20:25:00.216860] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09bd0 00:26:05.444 20:25:00 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:05.444 20:25:00 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:26:05.444 [2024-10-01 20:25:00.219523] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:26:06.083 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:06.083 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:06.083 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:06.083 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:06.083 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:06.083 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:06.083 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:06.083 20:25:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:06.083 20:25:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:06.083 20:25:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:06.083 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:06.083 "name": "raid_bdev1", 00:26:06.083 "uuid": "e463d538-8861-4ac1-82e4-8d298a9ec734", 00:26:06.083 "strip_size_kb": 0, 00:26:06.083 "state": "online", 00:26:06.083 "raid_level": "raid1", 00:26:06.083 "superblock": false, 00:26:06.083 "num_base_bdevs": 2, 00:26:06.083 "num_base_bdevs_discovered": 2, 00:26:06.083 "num_base_bdevs_operational": 2, 00:26:06.083 "process": { 00:26:06.083 "type": "rebuild", 00:26:06.083 "target": "spare", 00:26:06.083 "progress": { 00:26:06.083 "blocks": 20480, 00:26:06.083 "percent": 31 00:26:06.083 } 00:26:06.083 }, 00:26:06.083 "base_bdevs_list": [ 00:26:06.083 { 00:26:06.083 "name": "spare", 00:26:06.083 "uuid": "68d0b237-d012-5ff7-8705-93db54d4fd91", 00:26:06.083 "is_configured": true, 00:26:06.083 "data_offset": 0, 00:26:06.083 "data_size": 65536 00:26:06.083 }, 00:26:06.083 { 00:26:06.083 "name": "BaseBdev2", 00:26:06.083 "uuid": "4e9a1435-f39b-5e3f-8c2c-840c0d05770b", 00:26:06.083 "is_configured": true, 00:26:06.083 "data_offset": 0, 00:26:06.083 "data_size": 65536 00:26:06.083 } 00:26:06.083 ] 00:26:06.083 }' 00:26:06.083 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:06.342 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:06.342 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:06.342 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:26:06.342 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:26:06.342 20:25:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:06.342 20:25:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:06.342 [2024-10-01 20:25:01.412780] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:06.342 [2024-10-01 20:25:01.428770] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:26:06.342 [2024-10-01 20:25:01.428858] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:06.342 [2024-10-01 20:25:01.428882] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:06.342 [2024-10-01 20:25:01.428898] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:26:06.342 20:25:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:06.342 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:26:06.342 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:06.342 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:06.342 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:06.342 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:06.342 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:26:06.342 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:06.342 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:06.342 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:06.343 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:06.343 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:06.343 20:25:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:06.343 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:06.343 20:25:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:06.343 20:25:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:06.343 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:06.343 "name": "raid_bdev1", 00:26:06.343 "uuid": "e463d538-8861-4ac1-82e4-8d298a9ec734", 00:26:06.343 "strip_size_kb": 0, 00:26:06.343 "state": "online", 00:26:06.343 "raid_level": "raid1", 00:26:06.343 "superblock": false, 00:26:06.343 "num_base_bdevs": 2, 00:26:06.343 "num_base_bdevs_discovered": 1, 00:26:06.343 "num_base_bdevs_operational": 1, 00:26:06.343 "base_bdevs_list": [ 00:26:06.343 { 00:26:06.343 "name": null, 00:26:06.343 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:06.343 "is_configured": false, 00:26:06.343 "data_offset": 0, 00:26:06.343 "data_size": 65536 00:26:06.343 }, 00:26:06.343 { 00:26:06.343 "name": "BaseBdev2", 00:26:06.343 "uuid": "4e9a1435-f39b-5e3f-8c2c-840c0d05770b", 00:26:06.343 "is_configured": true, 00:26:06.343 "data_offset": 0, 00:26:06.343 "data_size": 65536 00:26:06.343 } 00:26:06.343 ] 00:26:06.343 }' 00:26:06.343 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:06.343 20:25:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:06.911 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:26:06.911 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:06.911 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:26:06.911 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:26:06.911 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:06.911 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:06.911 20:25:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:06.911 20:25:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:06.911 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:06.911 20:25:01 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:06.911 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:06.911 "name": "raid_bdev1", 00:26:06.911 "uuid": "e463d538-8861-4ac1-82e4-8d298a9ec734", 00:26:06.911 "strip_size_kb": 0, 00:26:06.911 "state": "online", 00:26:06.911 "raid_level": "raid1", 00:26:06.911 "superblock": false, 00:26:06.911 "num_base_bdevs": 2, 00:26:06.911 "num_base_bdevs_discovered": 1, 00:26:06.911 "num_base_bdevs_operational": 1, 00:26:06.911 "base_bdevs_list": [ 00:26:06.911 { 00:26:06.911 "name": null, 00:26:06.911 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:06.911 "is_configured": false, 00:26:06.911 "data_offset": 0, 00:26:06.911 "data_size": 65536 00:26:06.911 }, 00:26:06.911 { 00:26:06.911 "name": "BaseBdev2", 00:26:06.911 "uuid": "4e9a1435-f39b-5e3f-8c2c-840c0d05770b", 00:26:06.911 "is_configured": true, 00:26:06.911 "data_offset": 0, 00:26:06.911 "data_size": 65536 00:26:06.911 } 00:26:06.911 ] 00:26:06.911 }' 00:26:06.911 20:25:01 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:06.911 20:25:02 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:26:06.911 20:25:02 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:06.911 20:25:02 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:26:06.911 20:25:02 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:26:06.911 20:25:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:06.911 20:25:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:06.911 [2024-10-01 20:25:02.103068] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:26:06.911 [2024-10-01 20:25:02.117997] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09ca0 00:26:06.911 20:25:02 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:06.911 20:25:02 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:26:06.911 [2024-10-01 20:25:02.120678] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:26:08.286 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:08.286 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:08.286 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:08.286 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:08.286 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:08.286 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:08.286 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:08.286 20:25:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:08.286 20:25:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:08.286 20:25:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:08.286 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:08.286 "name": "raid_bdev1", 00:26:08.286 "uuid": "e463d538-8861-4ac1-82e4-8d298a9ec734", 00:26:08.286 "strip_size_kb": 0, 00:26:08.286 "state": "online", 00:26:08.286 "raid_level": "raid1", 00:26:08.286 "superblock": false, 00:26:08.286 "num_base_bdevs": 2, 00:26:08.286 "num_base_bdevs_discovered": 2, 00:26:08.286 "num_base_bdevs_operational": 2, 00:26:08.286 "process": { 00:26:08.286 "type": "rebuild", 00:26:08.286 "target": "spare", 00:26:08.286 "progress": { 00:26:08.286 "blocks": 20480, 00:26:08.286 "percent": 31 00:26:08.286 } 00:26:08.286 }, 00:26:08.286 "base_bdevs_list": [ 00:26:08.286 { 00:26:08.286 "name": "spare", 00:26:08.286 "uuid": "68d0b237-d012-5ff7-8705-93db54d4fd91", 00:26:08.286 "is_configured": true, 00:26:08.286 "data_offset": 0, 00:26:08.286 "data_size": 65536 00:26:08.286 }, 00:26:08.286 { 00:26:08.286 "name": "BaseBdev2", 00:26:08.286 "uuid": "4e9a1435-f39b-5e3f-8c2c-840c0d05770b", 00:26:08.286 "is_configured": true, 00:26:08.286 "data_offset": 0, 00:26:08.286 "data_size": 65536 00:26:08.286 } 00:26:08.286 ] 00:26:08.286 }' 00:26:08.286 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:08.286 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:08.286 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:08.286 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:26:08.286 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=442 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:08.287 "name": "raid_bdev1", 00:26:08.287 "uuid": "e463d538-8861-4ac1-82e4-8d298a9ec734", 00:26:08.287 "strip_size_kb": 0, 00:26:08.287 "state": "online", 00:26:08.287 "raid_level": "raid1", 00:26:08.287 "superblock": false, 00:26:08.287 "num_base_bdevs": 2, 00:26:08.287 "num_base_bdevs_discovered": 2, 00:26:08.287 "num_base_bdevs_operational": 2, 00:26:08.287 "process": { 00:26:08.287 "type": "rebuild", 00:26:08.287 "target": "spare", 00:26:08.287 "progress": { 00:26:08.287 "blocks": 22528, 00:26:08.287 "percent": 34 00:26:08.287 } 00:26:08.287 }, 00:26:08.287 "base_bdevs_list": [ 00:26:08.287 { 00:26:08.287 "name": "spare", 00:26:08.287 "uuid": "68d0b237-d012-5ff7-8705-93db54d4fd91", 00:26:08.287 "is_configured": true, 00:26:08.287 "data_offset": 0, 00:26:08.287 "data_size": 65536 00:26:08.287 }, 00:26:08.287 { 00:26:08.287 "name": "BaseBdev2", 00:26:08.287 "uuid": "4e9a1435-f39b-5e3f-8c2c-840c0d05770b", 00:26:08.287 "is_configured": true, 00:26:08.287 "data_offset": 0, 00:26:08.287 "data_size": 65536 00:26:08.287 } 00:26:08.287 ] 00:26:08.287 }' 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:26:08.287 20:25:03 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:26:09.223 20:25:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:26:09.223 20:25:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:09.223 20:25:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:09.223 20:25:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:09.223 20:25:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:09.223 20:25:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:09.223 20:25:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:09.223 20:25:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:09.223 20:25:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:09.223 20:25:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:09.223 20:25:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:09.482 20:25:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:09.482 "name": "raid_bdev1", 00:26:09.482 "uuid": "e463d538-8861-4ac1-82e4-8d298a9ec734", 00:26:09.482 "strip_size_kb": 0, 00:26:09.482 "state": "online", 00:26:09.482 "raid_level": "raid1", 00:26:09.482 "superblock": false, 00:26:09.482 "num_base_bdevs": 2, 00:26:09.482 "num_base_bdevs_discovered": 2, 00:26:09.482 "num_base_bdevs_operational": 2, 00:26:09.482 "process": { 00:26:09.482 "type": "rebuild", 00:26:09.482 "target": "spare", 00:26:09.482 "progress": { 00:26:09.482 "blocks": 47104, 00:26:09.482 "percent": 71 00:26:09.482 } 00:26:09.482 }, 00:26:09.482 "base_bdevs_list": [ 00:26:09.482 { 00:26:09.482 "name": "spare", 00:26:09.482 "uuid": "68d0b237-d012-5ff7-8705-93db54d4fd91", 00:26:09.482 "is_configured": true, 00:26:09.482 "data_offset": 0, 00:26:09.482 "data_size": 65536 00:26:09.482 }, 00:26:09.482 { 00:26:09.482 "name": "BaseBdev2", 00:26:09.482 "uuid": "4e9a1435-f39b-5e3f-8c2c-840c0d05770b", 00:26:09.482 "is_configured": true, 00:26:09.482 "data_offset": 0, 00:26:09.482 "data_size": 65536 00:26:09.482 } 00:26:09.482 ] 00:26:09.482 }' 00:26:09.482 20:25:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:09.482 20:25:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:09.482 20:25:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:09.482 20:25:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:26:09.482 20:25:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:26:10.416 [2024-10-01 20:25:05.345061] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:26:10.416 [2024-10-01 20:25:05.345205] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:26:10.416 [2024-10-01 20:25:05.345303] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:10.416 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:26:10.416 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:10.416 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:10.416 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:10.416 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:10.416 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:10.416 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:10.416 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:10.416 20:25:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:10.416 20:25:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:10.416 20:25:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:10.416 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:10.416 "name": "raid_bdev1", 00:26:10.416 "uuid": "e463d538-8861-4ac1-82e4-8d298a9ec734", 00:26:10.416 "strip_size_kb": 0, 00:26:10.416 "state": "online", 00:26:10.416 "raid_level": "raid1", 00:26:10.416 "superblock": false, 00:26:10.416 "num_base_bdevs": 2, 00:26:10.416 "num_base_bdevs_discovered": 2, 00:26:10.416 "num_base_bdevs_operational": 2, 00:26:10.416 "base_bdevs_list": [ 00:26:10.416 { 00:26:10.416 "name": "spare", 00:26:10.416 "uuid": "68d0b237-d012-5ff7-8705-93db54d4fd91", 00:26:10.416 "is_configured": true, 00:26:10.416 "data_offset": 0, 00:26:10.416 "data_size": 65536 00:26:10.416 }, 00:26:10.416 { 00:26:10.416 "name": "BaseBdev2", 00:26:10.416 "uuid": "4e9a1435-f39b-5e3f-8c2c-840c0d05770b", 00:26:10.416 "is_configured": true, 00:26:10.416 "data_offset": 0, 00:26:10.416 "data_size": 65536 00:26:10.416 } 00:26:10.416 ] 00:26:10.416 }' 00:26:10.416 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:10.674 "name": "raid_bdev1", 00:26:10.674 "uuid": "e463d538-8861-4ac1-82e4-8d298a9ec734", 00:26:10.674 "strip_size_kb": 0, 00:26:10.674 "state": "online", 00:26:10.674 "raid_level": "raid1", 00:26:10.674 "superblock": false, 00:26:10.674 "num_base_bdevs": 2, 00:26:10.674 "num_base_bdevs_discovered": 2, 00:26:10.674 "num_base_bdevs_operational": 2, 00:26:10.674 "base_bdevs_list": [ 00:26:10.674 { 00:26:10.674 "name": "spare", 00:26:10.674 "uuid": "68d0b237-d012-5ff7-8705-93db54d4fd91", 00:26:10.674 "is_configured": true, 00:26:10.674 "data_offset": 0, 00:26:10.674 "data_size": 65536 00:26:10.674 }, 00:26:10.674 { 00:26:10.674 "name": "BaseBdev2", 00:26:10.674 "uuid": "4e9a1435-f39b-5e3f-8c2c-840c0d05770b", 00:26:10.674 "is_configured": true, 00:26:10.674 "data_offset": 0, 00:26:10.674 "data_size": 65536 00:26:10.674 } 00:26:10.674 ] 00:26:10.674 }' 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:10.674 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:10.932 20:25:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:10.932 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:10.932 "name": "raid_bdev1", 00:26:10.932 "uuid": "e463d538-8861-4ac1-82e4-8d298a9ec734", 00:26:10.932 "strip_size_kb": 0, 00:26:10.932 "state": "online", 00:26:10.932 "raid_level": "raid1", 00:26:10.932 "superblock": false, 00:26:10.932 "num_base_bdevs": 2, 00:26:10.932 "num_base_bdevs_discovered": 2, 00:26:10.932 "num_base_bdevs_operational": 2, 00:26:10.932 "base_bdevs_list": [ 00:26:10.932 { 00:26:10.932 "name": "spare", 00:26:10.932 "uuid": "68d0b237-d012-5ff7-8705-93db54d4fd91", 00:26:10.932 "is_configured": true, 00:26:10.932 "data_offset": 0, 00:26:10.932 "data_size": 65536 00:26:10.932 }, 00:26:10.932 { 00:26:10.932 "name": "BaseBdev2", 00:26:10.932 "uuid": "4e9a1435-f39b-5e3f-8c2c-840c0d05770b", 00:26:10.932 "is_configured": true, 00:26:10.932 "data_offset": 0, 00:26:10.932 "data_size": 65536 00:26:10.932 } 00:26:10.932 ] 00:26:10.932 }' 00:26:10.932 20:25:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:10.932 20:25:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:11.190 20:25:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:26:11.190 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:11.190 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:11.190 [2024-10-01 20:25:06.431404] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:26:11.190 [2024-10-01 20:25:06.431452] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:26:11.190 [2024-10-01 20:25:06.431580] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:11.190 [2024-10-01 20:25:06.431694] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:11.190 [2024-10-01 20:25:06.431753] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:26:11.190 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:11.190 20:25:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:11.190 20:25:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:26:11.190 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:11.190 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:11.448 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:11.448 20:25:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:26:11.448 20:25:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:26:11.448 20:25:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:26:11.449 20:25:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:26:11.449 20:25:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:26:11.449 20:25:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:26:11.449 20:25:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:26:11.449 20:25:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:26:11.449 20:25:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:26:11.449 20:25:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:26:11.449 20:25:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:26:11.449 20:25:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:26:11.449 20:25:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:26:11.707 /dev/nbd0 00:26:11.707 20:25:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:26:11.707 20:25:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:26:11.707 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:26:11.707 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:26:11.707 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:26:11.707 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:26:11.707 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:26:11.707 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:26:11.707 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:26:11.707 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:26:11.707 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:26:11.708 1+0 records in 00:26:11.708 1+0 records out 00:26:11.708 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000309704 s, 13.2 MB/s 00:26:11.708 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:11.708 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:26:11.708 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:11.708 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:26:11.708 20:25:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:26:11.708 20:25:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:26:11.708 20:25:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:26:11.708 20:25:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:26:11.966 /dev/nbd1 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:26:11.966 1+0 records in 00:26:11.966 1+0 records out 00:26:11.966 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000399537 s, 10.3 MB/s 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:26:11.966 20:25:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:26:12.224 20:25:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:26:12.224 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:26:12.224 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:26:12.224 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:26:12.224 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:26:12.224 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:26:12.224 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:26:12.483 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:26:12.483 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:26:12.483 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:26:12.483 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:26:12.483 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:26:12.483 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:26:12.483 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:26:12.483 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:26:12.483 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:26:12.483 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:26:12.741 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:26:12.741 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:26:12.741 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 76217 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@950 -- # '[' -z 76217 ']' 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # kill -0 76217 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@955 -- # uname 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 76217 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:26:12.742 killing process with pid 76217 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 76217' 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@969 -- # kill 76217 00:26:12.742 Received shutdown signal, test time was about 60.000000 seconds 00:26:12.742 00:26:12.742 Latency(us) 00:26:12.742 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:12.742 =================================================================================================================== 00:26:12.742 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:26:12.742 [2024-10-01 20:25:07.897308] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:12.742 20:25:07 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@974 -- # wait 76217 00:26:13.000 [2024-10-01 20:25:08.165601] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:15.031 20:25:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:26:15.031 00:26:15.031 real 0m19.928s 00:26:15.031 user 0m21.958s 00:26:15.031 sys 0m3.673s 00:26:15.031 20:25:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:26:15.031 20:25:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:26:15.031 ************************************ 00:26:15.031 END TEST raid_rebuild_test 00:26:15.031 ************************************ 00:26:15.031 20:25:10 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 2 true false true 00:26:15.031 20:25:10 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:26:15.031 20:25:10 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:26:15.031 20:25:10 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:15.031 ************************************ 00:26:15.031 START TEST raid_rebuild_test_sb 00:26:15.031 ************************************ 00:26:15.031 20:25:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true false true 00:26:15.031 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:26:15.031 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=76682 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 76682 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@831 -- # '[' -z 76682 ']' 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:26:15.032 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:26:15.032 20:25:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:15.032 [2024-10-01 20:25:10.166139] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:26:15.032 I/O size of 3145728 is greater than zero copy threshold (65536). 00:26:15.032 Zero copy mechanism will not be used. 00:26:15.032 [2024-10-01 20:25:10.166358] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76682 ] 00:26:15.291 [2024-10-01 20:25:10.336856] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:15.549 [2024-10-01 20:25:10.589072] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:26:15.549 [2024-10-01 20:25:10.795389] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:15.549 [2024-10-01 20:25:10.795432] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # return 0 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:16.117 BaseBdev1_malloc 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:16.117 [2024-10-01 20:25:11.259351] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:26:16.117 [2024-10-01 20:25:11.259439] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:16.117 [2024-10-01 20:25:11.259477] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:26:16.117 [2024-10-01 20:25:11.259503] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:16.117 [2024-10-01 20:25:11.262455] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:16.117 [2024-10-01 20:25:11.262645] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:26:16.117 BaseBdev1 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:16.117 BaseBdev2_malloc 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:16.117 [2024-10-01 20:25:11.316245] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:26:16.117 [2024-10-01 20:25:11.316462] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:16.117 [2024-10-01 20:25:11.316511] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:26:16.117 [2024-10-01 20:25:11.316532] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:16.117 [2024-10-01 20:25:11.319408] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:16.117 [2024-10-01 20:25:11.319464] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:26:16.117 BaseBdev2 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:16.117 spare_malloc 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:16.117 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:16.377 spare_delay 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:16.377 [2024-10-01 20:25:11.380993] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:26:16.377 [2024-10-01 20:25:11.381079] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:16.377 [2024-10-01 20:25:11.381111] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:26:16.377 [2024-10-01 20:25:11.381132] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:16.377 [2024-10-01 20:25:11.384044] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:16.377 [2024-10-01 20:25:11.384100] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:26:16.377 spare 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:16.377 [2024-10-01 20:25:11.389098] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:16.377 [2024-10-01 20:25:11.391626] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:26:16.377 [2024-10-01 20:25:11.392043] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:26:16.377 [2024-10-01 20:25:11.392077] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:26:16.377 [2024-10-01 20:25:11.392417] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:26:16.377 [2024-10-01 20:25:11.392639] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:26:16.377 [2024-10-01 20:25:11.392657] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:26:16.377 [2024-10-01 20:25:11.392866] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:16.377 "name": "raid_bdev1", 00:26:16.377 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:16.377 "strip_size_kb": 0, 00:26:16.377 "state": "online", 00:26:16.377 "raid_level": "raid1", 00:26:16.377 "superblock": true, 00:26:16.377 "num_base_bdevs": 2, 00:26:16.377 "num_base_bdevs_discovered": 2, 00:26:16.377 "num_base_bdevs_operational": 2, 00:26:16.377 "base_bdevs_list": [ 00:26:16.377 { 00:26:16.377 "name": "BaseBdev1", 00:26:16.377 "uuid": "67a69c51-28a1-588f-93b8-675fff7e28da", 00:26:16.377 "is_configured": true, 00:26:16.377 "data_offset": 2048, 00:26:16.377 "data_size": 63488 00:26:16.377 }, 00:26:16.377 { 00:26:16.377 "name": "BaseBdev2", 00:26:16.377 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:16.377 "is_configured": true, 00:26:16.377 "data_offset": 2048, 00:26:16.377 "data_size": 63488 00:26:16.377 } 00:26:16.377 ] 00:26:16.377 }' 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:16.377 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:16.637 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:26:16.637 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:16.637 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:16.637 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:26:16.637 [2024-10-01 20:25:11.889621] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:16.896 20:25:11 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:26:17.155 [2024-10-01 20:25:12.269450] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:26:17.155 /dev/nbd0 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:26:17.155 1+0 records in 00:26:17.155 1+0 records out 00:26:17.155 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000383586 s, 10.7 MB/s 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:26:17.155 20:25:12 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:26:23.720 63488+0 records in 00:26:23.720 63488+0 records out 00:26:23.720 32505856 bytes (33 MB, 31 MiB) copied, 6.37355 s, 5.1 MB/s 00:26:23.720 20:25:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:26:23.720 20:25:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:26:23.720 20:25:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:26:23.720 20:25:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:26:23.720 20:25:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:26:23.720 20:25:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:26:23.720 20:25:18 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:26:23.978 [2024-10-01 20:25:19.006176] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:23.978 [2024-10-01 20:25:19.024314] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:23.978 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:23.979 "name": "raid_bdev1", 00:26:23.979 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:23.979 "strip_size_kb": 0, 00:26:23.979 "state": "online", 00:26:23.979 "raid_level": "raid1", 00:26:23.979 "superblock": true, 00:26:23.979 "num_base_bdevs": 2, 00:26:23.979 "num_base_bdevs_discovered": 1, 00:26:23.979 "num_base_bdevs_operational": 1, 00:26:23.979 "base_bdevs_list": [ 00:26:23.979 { 00:26:23.979 "name": null, 00:26:23.979 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:23.979 "is_configured": false, 00:26:23.979 "data_offset": 0, 00:26:23.979 "data_size": 63488 00:26:23.979 }, 00:26:23.979 { 00:26:23.979 "name": "BaseBdev2", 00:26:23.979 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:23.979 "is_configured": true, 00:26:23.979 "data_offset": 2048, 00:26:23.979 "data_size": 63488 00:26:23.979 } 00:26:23.979 ] 00:26:23.979 }' 00:26:23.979 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:23.979 20:25:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:24.546 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:26:24.546 20:25:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:24.546 20:25:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:24.546 [2024-10-01 20:25:19.560513] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:26:24.546 [2024-10-01 20:25:19.576572] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3360 00:26:24.546 20:25:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:24.546 20:25:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:26:24.546 [2024-10-01 20:25:19.579233] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:26:25.480 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:25.480 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:25.480 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:25.480 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:25.481 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:25.481 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:25.481 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:25.481 20:25:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:25.481 20:25:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:25.481 20:25:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:25.481 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:25.481 "name": "raid_bdev1", 00:26:25.481 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:25.481 "strip_size_kb": 0, 00:26:25.481 "state": "online", 00:26:25.481 "raid_level": "raid1", 00:26:25.481 "superblock": true, 00:26:25.481 "num_base_bdevs": 2, 00:26:25.481 "num_base_bdevs_discovered": 2, 00:26:25.481 "num_base_bdevs_operational": 2, 00:26:25.481 "process": { 00:26:25.481 "type": "rebuild", 00:26:25.481 "target": "spare", 00:26:25.481 "progress": { 00:26:25.481 "blocks": 20480, 00:26:25.481 "percent": 32 00:26:25.481 } 00:26:25.481 }, 00:26:25.481 "base_bdevs_list": [ 00:26:25.481 { 00:26:25.481 "name": "spare", 00:26:25.481 "uuid": "db533bec-185a-5cf6-84ec-12fc4f29dabf", 00:26:25.481 "is_configured": true, 00:26:25.481 "data_offset": 2048, 00:26:25.481 "data_size": 63488 00:26:25.481 }, 00:26:25.481 { 00:26:25.481 "name": "BaseBdev2", 00:26:25.481 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:25.481 "is_configured": true, 00:26:25.481 "data_offset": 2048, 00:26:25.481 "data_size": 63488 00:26:25.481 } 00:26:25.481 ] 00:26:25.481 }' 00:26:25.481 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:25.481 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:25.481 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:25.739 [2024-10-01 20:25:20.752565] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:25.739 [2024-10-01 20:25:20.788552] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:26:25.739 [2024-10-01 20:25:20.788692] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:25.739 [2024-10-01 20:25:20.788717] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:25.739 [2024-10-01 20:25:20.788762] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:25.739 "name": "raid_bdev1", 00:26:25.739 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:25.739 "strip_size_kb": 0, 00:26:25.739 "state": "online", 00:26:25.739 "raid_level": "raid1", 00:26:25.739 "superblock": true, 00:26:25.739 "num_base_bdevs": 2, 00:26:25.739 "num_base_bdevs_discovered": 1, 00:26:25.739 "num_base_bdevs_operational": 1, 00:26:25.739 "base_bdevs_list": [ 00:26:25.739 { 00:26:25.739 "name": null, 00:26:25.739 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:25.739 "is_configured": false, 00:26:25.739 "data_offset": 0, 00:26:25.739 "data_size": 63488 00:26:25.739 }, 00:26:25.739 { 00:26:25.739 "name": "BaseBdev2", 00:26:25.739 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:25.739 "is_configured": true, 00:26:25.739 "data_offset": 2048, 00:26:25.739 "data_size": 63488 00:26:25.739 } 00:26:25.739 ] 00:26:25.739 }' 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:25.739 20:25:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:26.306 "name": "raid_bdev1", 00:26:26.306 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:26.306 "strip_size_kb": 0, 00:26:26.306 "state": "online", 00:26:26.306 "raid_level": "raid1", 00:26:26.306 "superblock": true, 00:26:26.306 "num_base_bdevs": 2, 00:26:26.306 "num_base_bdevs_discovered": 1, 00:26:26.306 "num_base_bdevs_operational": 1, 00:26:26.306 "base_bdevs_list": [ 00:26:26.306 { 00:26:26.306 "name": null, 00:26:26.306 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:26.306 "is_configured": false, 00:26:26.306 "data_offset": 0, 00:26:26.306 "data_size": 63488 00:26:26.306 }, 00:26:26.306 { 00:26:26.306 "name": "BaseBdev2", 00:26:26.306 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:26.306 "is_configured": true, 00:26:26.306 "data_offset": 2048, 00:26:26.306 "data_size": 63488 00:26:26.306 } 00:26:26.306 ] 00:26:26.306 }' 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:26.306 [2024-10-01 20:25:21.490869] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:26:26.306 [2024-10-01 20:25:21.506132] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3430 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:26.306 20:25:21 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:26:26.306 [2024-10-01 20:25:21.508770] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:27.690 "name": "raid_bdev1", 00:26:27.690 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:27.690 "strip_size_kb": 0, 00:26:27.690 "state": "online", 00:26:27.690 "raid_level": "raid1", 00:26:27.690 "superblock": true, 00:26:27.690 "num_base_bdevs": 2, 00:26:27.690 "num_base_bdevs_discovered": 2, 00:26:27.690 "num_base_bdevs_operational": 2, 00:26:27.690 "process": { 00:26:27.690 "type": "rebuild", 00:26:27.690 "target": "spare", 00:26:27.690 "progress": { 00:26:27.690 "blocks": 20480, 00:26:27.690 "percent": 32 00:26:27.690 } 00:26:27.690 }, 00:26:27.690 "base_bdevs_list": [ 00:26:27.690 { 00:26:27.690 "name": "spare", 00:26:27.690 "uuid": "db533bec-185a-5cf6-84ec-12fc4f29dabf", 00:26:27.690 "is_configured": true, 00:26:27.690 "data_offset": 2048, 00:26:27.690 "data_size": 63488 00:26:27.690 }, 00:26:27.690 { 00:26:27.690 "name": "BaseBdev2", 00:26:27.690 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:27.690 "is_configured": true, 00:26:27.690 "data_offset": 2048, 00:26:27.690 "data_size": 63488 00:26:27.690 } 00:26:27.690 ] 00:26:27.690 }' 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:26:27.690 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=461 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:27.690 "name": "raid_bdev1", 00:26:27.690 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:27.690 "strip_size_kb": 0, 00:26:27.690 "state": "online", 00:26:27.690 "raid_level": "raid1", 00:26:27.690 "superblock": true, 00:26:27.690 "num_base_bdevs": 2, 00:26:27.690 "num_base_bdevs_discovered": 2, 00:26:27.690 "num_base_bdevs_operational": 2, 00:26:27.690 "process": { 00:26:27.690 "type": "rebuild", 00:26:27.690 "target": "spare", 00:26:27.690 "progress": { 00:26:27.690 "blocks": 22528, 00:26:27.690 "percent": 35 00:26:27.690 } 00:26:27.690 }, 00:26:27.690 "base_bdevs_list": [ 00:26:27.690 { 00:26:27.690 "name": "spare", 00:26:27.690 "uuid": "db533bec-185a-5cf6-84ec-12fc4f29dabf", 00:26:27.690 "is_configured": true, 00:26:27.690 "data_offset": 2048, 00:26:27.690 "data_size": 63488 00:26:27.690 }, 00:26:27.690 { 00:26:27.690 "name": "BaseBdev2", 00:26:27.690 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:27.690 "is_configured": true, 00:26:27.690 "data_offset": 2048, 00:26:27.690 "data_size": 63488 00:26:27.690 } 00:26:27.690 ] 00:26:27.690 }' 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:26:27.690 20:25:22 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:26:28.659 20:25:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:26:28.659 20:25:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:28.659 20:25:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:28.659 20:25:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:28.659 20:25:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:28.659 20:25:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:28.659 20:25:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:28.659 20:25:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:28.659 20:25:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:28.659 20:25:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:28.659 20:25:23 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:28.659 20:25:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:28.659 "name": "raid_bdev1", 00:26:28.659 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:28.659 "strip_size_kb": 0, 00:26:28.659 "state": "online", 00:26:28.659 "raid_level": "raid1", 00:26:28.659 "superblock": true, 00:26:28.659 "num_base_bdevs": 2, 00:26:28.659 "num_base_bdevs_discovered": 2, 00:26:28.659 "num_base_bdevs_operational": 2, 00:26:28.659 "process": { 00:26:28.659 "type": "rebuild", 00:26:28.659 "target": "spare", 00:26:28.659 "progress": { 00:26:28.659 "blocks": 47104, 00:26:28.659 "percent": 74 00:26:28.659 } 00:26:28.659 }, 00:26:28.659 "base_bdevs_list": [ 00:26:28.659 { 00:26:28.659 "name": "spare", 00:26:28.659 "uuid": "db533bec-185a-5cf6-84ec-12fc4f29dabf", 00:26:28.659 "is_configured": true, 00:26:28.659 "data_offset": 2048, 00:26:28.659 "data_size": 63488 00:26:28.659 }, 00:26:28.659 { 00:26:28.659 "name": "BaseBdev2", 00:26:28.659 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:28.659 "is_configured": true, 00:26:28.659 "data_offset": 2048, 00:26:28.659 "data_size": 63488 00:26:28.659 } 00:26:28.659 ] 00:26:28.659 }' 00:26:28.659 20:25:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:28.933 20:25:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:28.933 20:25:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:28.933 20:25:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:26:28.933 20:25:23 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:26:29.500 [2024-10-01 20:25:24.631681] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:26:29.500 [2024-10-01 20:25:24.631857] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:26:29.500 [2024-10-01 20:25:24.632005] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:29.758 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:26:29.758 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:29.758 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:29.758 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:29.758 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:29.758 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:29.758 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:29.758 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:29.758 20:25:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:29.758 20:25:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:30.018 "name": "raid_bdev1", 00:26:30.018 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:30.018 "strip_size_kb": 0, 00:26:30.018 "state": "online", 00:26:30.018 "raid_level": "raid1", 00:26:30.018 "superblock": true, 00:26:30.018 "num_base_bdevs": 2, 00:26:30.018 "num_base_bdevs_discovered": 2, 00:26:30.018 "num_base_bdevs_operational": 2, 00:26:30.018 "base_bdevs_list": [ 00:26:30.018 { 00:26:30.018 "name": "spare", 00:26:30.018 "uuid": "db533bec-185a-5cf6-84ec-12fc4f29dabf", 00:26:30.018 "is_configured": true, 00:26:30.018 "data_offset": 2048, 00:26:30.018 "data_size": 63488 00:26:30.018 }, 00:26:30.018 { 00:26:30.018 "name": "BaseBdev2", 00:26:30.018 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:30.018 "is_configured": true, 00:26:30.018 "data_offset": 2048, 00:26:30.018 "data_size": 63488 00:26:30.018 } 00:26:30.018 ] 00:26:30.018 }' 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:30.018 "name": "raid_bdev1", 00:26:30.018 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:30.018 "strip_size_kb": 0, 00:26:30.018 "state": "online", 00:26:30.018 "raid_level": "raid1", 00:26:30.018 "superblock": true, 00:26:30.018 "num_base_bdevs": 2, 00:26:30.018 "num_base_bdevs_discovered": 2, 00:26:30.018 "num_base_bdevs_operational": 2, 00:26:30.018 "base_bdevs_list": [ 00:26:30.018 { 00:26:30.018 "name": "spare", 00:26:30.018 "uuid": "db533bec-185a-5cf6-84ec-12fc4f29dabf", 00:26:30.018 "is_configured": true, 00:26:30.018 "data_offset": 2048, 00:26:30.018 "data_size": 63488 00:26:30.018 }, 00:26:30.018 { 00:26:30.018 "name": "BaseBdev2", 00:26:30.018 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:30.018 "is_configured": true, 00:26:30.018 "data_offset": 2048, 00:26:30.018 "data_size": 63488 00:26:30.018 } 00:26:30.018 ] 00:26:30.018 }' 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:26:30.018 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:30.277 "name": "raid_bdev1", 00:26:30.277 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:30.277 "strip_size_kb": 0, 00:26:30.277 "state": "online", 00:26:30.277 "raid_level": "raid1", 00:26:30.277 "superblock": true, 00:26:30.277 "num_base_bdevs": 2, 00:26:30.277 "num_base_bdevs_discovered": 2, 00:26:30.277 "num_base_bdevs_operational": 2, 00:26:30.277 "base_bdevs_list": [ 00:26:30.277 { 00:26:30.277 "name": "spare", 00:26:30.277 "uuid": "db533bec-185a-5cf6-84ec-12fc4f29dabf", 00:26:30.277 "is_configured": true, 00:26:30.277 "data_offset": 2048, 00:26:30.277 "data_size": 63488 00:26:30.277 }, 00:26:30.277 { 00:26:30.277 "name": "BaseBdev2", 00:26:30.277 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:30.277 "is_configured": true, 00:26:30.277 "data_offset": 2048, 00:26:30.277 "data_size": 63488 00:26:30.277 } 00:26:30.277 ] 00:26:30.277 }' 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:30.277 20:25:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:30.846 [2024-10-01 20:25:25.835003] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:26:30.846 [2024-10-01 20:25:25.835047] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:26:30.846 [2024-10-01 20:25:25.835165] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:30.846 [2024-10-01 20:25:25.835263] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:30.846 [2024-10-01 20:25:25.835282] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:26:30.846 20:25:25 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:26:31.105 /dev/nbd0 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:26:31.105 1+0 records in 00:26:31.105 1+0 records out 00:26:31.105 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000258195 s, 15.9 MB/s 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:26:31.105 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:26:31.364 /dev/nbd1 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:26:31.364 1+0 records in 00:26:31.364 1+0 records out 00:26:31.364 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000346428 s, 11.8 MB/s 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:26:31.364 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:26:31.627 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:26:31.627 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:26:31.627 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:26:31.627 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:26:31.628 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:26:31.628 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:26:31.628 20:25:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:26:31.891 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:26:31.891 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:26:31.891 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:26:31.891 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:26:31.891 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:26:31.891 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:26:31.891 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:26:31.891 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:26:31.891 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:26:31.891 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:32.459 [2024-10-01 20:25:27.439224] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:26:32.459 [2024-10-01 20:25:27.439298] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:32.459 [2024-10-01 20:25:27.439342] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:26:32.459 [2024-10-01 20:25:27.439360] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:32.459 [2024-10-01 20:25:27.442426] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:32.459 [2024-10-01 20:25:27.442479] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:26:32.459 [2024-10-01 20:25:27.442613] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:26:32.459 [2024-10-01 20:25:27.442681] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:26:32.459 [2024-10-01 20:25:27.442902] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:26:32.459 spare 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:32.459 [2024-10-01 20:25:27.543063] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:26:32.459 [2024-10-01 20:25:27.543150] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:26:32.459 [2024-10-01 20:25:27.543627] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1ae0 00:26:32.459 [2024-10-01 20:25:27.543960] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:26:32.459 [2024-10-01 20:25:27.543991] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:26:32.459 [2024-10-01 20:25:27.544270] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:32.459 "name": "raid_bdev1", 00:26:32.459 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:32.459 "strip_size_kb": 0, 00:26:32.459 "state": "online", 00:26:32.459 "raid_level": "raid1", 00:26:32.459 "superblock": true, 00:26:32.459 "num_base_bdevs": 2, 00:26:32.459 "num_base_bdevs_discovered": 2, 00:26:32.459 "num_base_bdevs_operational": 2, 00:26:32.459 "base_bdevs_list": [ 00:26:32.459 { 00:26:32.459 "name": "spare", 00:26:32.459 "uuid": "db533bec-185a-5cf6-84ec-12fc4f29dabf", 00:26:32.459 "is_configured": true, 00:26:32.459 "data_offset": 2048, 00:26:32.459 "data_size": 63488 00:26:32.459 }, 00:26:32.459 { 00:26:32.459 "name": "BaseBdev2", 00:26:32.459 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:32.459 "is_configured": true, 00:26:32.459 "data_offset": 2048, 00:26:32.459 "data_size": 63488 00:26:32.459 } 00:26:32.459 ] 00:26:32.459 }' 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:32.459 20:25:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:33.027 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:26:33.027 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:33.027 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:26:33.027 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:26:33.027 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:33.027 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:33.027 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:33.027 20:25:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:33.027 20:25:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:33.027 20:25:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:33.027 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:33.027 "name": "raid_bdev1", 00:26:33.027 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:33.027 "strip_size_kb": 0, 00:26:33.027 "state": "online", 00:26:33.027 "raid_level": "raid1", 00:26:33.027 "superblock": true, 00:26:33.027 "num_base_bdevs": 2, 00:26:33.027 "num_base_bdevs_discovered": 2, 00:26:33.027 "num_base_bdevs_operational": 2, 00:26:33.027 "base_bdevs_list": [ 00:26:33.027 { 00:26:33.027 "name": "spare", 00:26:33.027 "uuid": "db533bec-185a-5cf6-84ec-12fc4f29dabf", 00:26:33.027 "is_configured": true, 00:26:33.027 "data_offset": 2048, 00:26:33.027 "data_size": 63488 00:26:33.027 }, 00:26:33.027 { 00:26:33.027 "name": "BaseBdev2", 00:26:33.027 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:33.027 "is_configured": true, 00:26:33.027 "data_offset": 2048, 00:26:33.028 "data_size": 63488 00:26:33.028 } 00:26:33.028 ] 00:26:33.028 }' 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:33.028 [2024-10-01 20:25:28.268428] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:33.028 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:33.287 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:33.287 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:33.287 20:25:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:33.287 20:25:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:33.287 20:25:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:33.287 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:33.287 "name": "raid_bdev1", 00:26:33.287 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:33.287 "strip_size_kb": 0, 00:26:33.287 "state": "online", 00:26:33.287 "raid_level": "raid1", 00:26:33.287 "superblock": true, 00:26:33.287 "num_base_bdevs": 2, 00:26:33.287 "num_base_bdevs_discovered": 1, 00:26:33.287 "num_base_bdevs_operational": 1, 00:26:33.287 "base_bdevs_list": [ 00:26:33.287 { 00:26:33.287 "name": null, 00:26:33.287 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:33.287 "is_configured": false, 00:26:33.287 "data_offset": 0, 00:26:33.287 "data_size": 63488 00:26:33.287 }, 00:26:33.287 { 00:26:33.287 "name": "BaseBdev2", 00:26:33.287 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:33.287 "is_configured": true, 00:26:33.287 "data_offset": 2048, 00:26:33.287 "data_size": 63488 00:26:33.287 } 00:26:33.287 ] 00:26:33.287 }' 00:26:33.287 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:33.287 20:25:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:33.545 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:26:33.545 20:25:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:33.545 20:25:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:33.805 [2024-10-01 20:25:28.800626] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:26:33.805 [2024-10-01 20:25:28.800901] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:26:33.805 [2024-10-01 20:25:28.800932] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:26:33.805 [2024-10-01 20:25:28.800989] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:26:33.805 [2024-10-01 20:25:28.815502] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1bb0 00:26:33.805 20:25:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:33.805 20:25:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:26:33.805 [2024-10-01 20:25:28.818168] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:26:34.738 20:25:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:34.738 20:25:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:34.738 20:25:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:34.738 20:25:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:34.738 20:25:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:34.738 20:25:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:34.738 20:25:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:34.738 20:25:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:34.738 20:25:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:34.738 20:25:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:34.738 20:25:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:34.738 "name": "raid_bdev1", 00:26:34.738 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:34.738 "strip_size_kb": 0, 00:26:34.738 "state": "online", 00:26:34.738 "raid_level": "raid1", 00:26:34.738 "superblock": true, 00:26:34.738 "num_base_bdevs": 2, 00:26:34.738 "num_base_bdevs_discovered": 2, 00:26:34.738 "num_base_bdevs_operational": 2, 00:26:34.738 "process": { 00:26:34.738 "type": "rebuild", 00:26:34.738 "target": "spare", 00:26:34.738 "progress": { 00:26:34.738 "blocks": 20480, 00:26:34.738 "percent": 32 00:26:34.738 } 00:26:34.738 }, 00:26:34.738 "base_bdevs_list": [ 00:26:34.738 { 00:26:34.738 "name": "spare", 00:26:34.738 "uuid": "db533bec-185a-5cf6-84ec-12fc4f29dabf", 00:26:34.738 "is_configured": true, 00:26:34.738 "data_offset": 2048, 00:26:34.738 "data_size": 63488 00:26:34.738 }, 00:26:34.738 { 00:26:34.738 "name": "BaseBdev2", 00:26:34.738 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:34.738 "is_configured": true, 00:26:34.738 "data_offset": 2048, 00:26:34.738 "data_size": 63488 00:26:34.738 } 00:26:34.738 ] 00:26:34.738 }' 00:26:34.738 20:25:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:34.738 20:25:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:34.738 20:25:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:34.996 20:25:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:26:34.996 20:25:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:26:34.996 20:25:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:34.996 20:25:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:34.996 [2024-10-01 20:25:29.999594] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:34.996 [2024-10-01 20:25:30.027483] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:26:34.996 [2024-10-01 20:25:30.027598] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:34.996 [2024-10-01 20:25:30.027624] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:34.996 [2024-10-01 20:25:30.027641] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:34.996 20:25:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:34.996 "name": "raid_bdev1", 00:26:34.997 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:34.997 "strip_size_kb": 0, 00:26:34.997 "state": "online", 00:26:34.997 "raid_level": "raid1", 00:26:34.997 "superblock": true, 00:26:34.997 "num_base_bdevs": 2, 00:26:34.997 "num_base_bdevs_discovered": 1, 00:26:34.997 "num_base_bdevs_operational": 1, 00:26:34.997 "base_bdevs_list": [ 00:26:34.997 { 00:26:34.997 "name": null, 00:26:34.997 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:34.997 "is_configured": false, 00:26:34.997 "data_offset": 0, 00:26:34.997 "data_size": 63488 00:26:34.997 }, 00:26:34.997 { 00:26:34.997 "name": "BaseBdev2", 00:26:34.997 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:34.997 "is_configured": true, 00:26:34.997 "data_offset": 2048, 00:26:34.997 "data_size": 63488 00:26:34.997 } 00:26:34.997 ] 00:26:34.997 }' 00:26:34.997 20:25:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:34.997 20:25:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:35.563 20:25:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:26:35.563 20:25:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:35.563 20:25:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:35.563 [2024-10-01 20:25:30.581642] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:26:35.563 [2024-10-01 20:25:30.581752] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:35.563 [2024-10-01 20:25:30.581792] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:26:35.563 [2024-10-01 20:25:30.581813] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:35.563 [2024-10-01 20:25:30.582453] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:35.563 [2024-10-01 20:25:30.582501] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:26:35.563 [2024-10-01 20:25:30.582633] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:26:35.563 [2024-10-01 20:25:30.582675] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:26:35.563 [2024-10-01 20:25:30.582699] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:26:35.563 [2024-10-01 20:25:30.582753] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:26:35.563 [2024-10-01 20:25:30.597479] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1c80 00:26:35.563 spare 00:26:35.563 20:25:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:35.563 20:25:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:26:35.563 [2024-10-01 20:25:30.600074] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:26:36.496 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:36.496 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:36.496 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:36.496 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:36.496 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:36.496 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:36.496 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:36.496 20:25:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:36.496 20:25:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:36.496 20:25:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:36.496 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:36.496 "name": "raid_bdev1", 00:26:36.496 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:36.496 "strip_size_kb": 0, 00:26:36.496 "state": "online", 00:26:36.496 "raid_level": "raid1", 00:26:36.496 "superblock": true, 00:26:36.496 "num_base_bdevs": 2, 00:26:36.496 "num_base_bdevs_discovered": 2, 00:26:36.496 "num_base_bdevs_operational": 2, 00:26:36.496 "process": { 00:26:36.496 "type": "rebuild", 00:26:36.496 "target": "spare", 00:26:36.496 "progress": { 00:26:36.496 "blocks": 20480, 00:26:36.496 "percent": 32 00:26:36.496 } 00:26:36.496 }, 00:26:36.496 "base_bdevs_list": [ 00:26:36.496 { 00:26:36.496 "name": "spare", 00:26:36.496 "uuid": "db533bec-185a-5cf6-84ec-12fc4f29dabf", 00:26:36.496 "is_configured": true, 00:26:36.496 "data_offset": 2048, 00:26:36.496 "data_size": 63488 00:26:36.496 }, 00:26:36.496 { 00:26:36.496 "name": "BaseBdev2", 00:26:36.496 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:36.496 "is_configured": true, 00:26:36.496 "data_offset": 2048, 00:26:36.496 "data_size": 63488 00:26:36.496 } 00:26:36.496 ] 00:26:36.496 }' 00:26:36.496 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:36.496 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:36.496 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:36.754 [2024-10-01 20:25:31.773466] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:36.754 [2024-10-01 20:25:31.809370] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:26:36.754 [2024-10-01 20:25:31.809469] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:36.754 [2024-10-01 20:25:31.809500] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:36.754 [2024-10-01 20:25:31.809513] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:36.754 "name": "raid_bdev1", 00:26:36.754 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:36.754 "strip_size_kb": 0, 00:26:36.754 "state": "online", 00:26:36.754 "raid_level": "raid1", 00:26:36.754 "superblock": true, 00:26:36.754 "num_base_bdevs": 2, 00:26:36.754 "num_base_bdevs_discovered": 1, 00:26:36.754 "num_base_bdevs_operational": 1, 00:26:36.754 "base_bdevs_list": [ 00:26:36.754 { 00:26:36.754 "name": null, 00:26:36.754 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:36.754 "is_configured": false, 00:26:36.754 "data_offset": 0, 00:26:36.754 "data_size": 63488 00:26:36.754 }, 00:26:36.754 { 00:26:36.754 "name": "BaseBdev2", 00:26:36.754 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:36.754 "is_configured": true, 00:26:36.754 "data_offset": 2048, 00:26:36.754 "data_size": 63488 00:26:36.754 } 00:26:36.754 ] 00:26:36.754 }' 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:36.754 20:25:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:37.321 "name": "raid_bdev1", 00:26:37.321 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:37.321 "strip_size_kb": 0, 00:26:37.321 "state": "online", 00:26:37.321 "raid_level": "raid1", 00:26:37.321 "superblock": true, 00:26:37.321 "num_base_bdevs": 2, 00:26:37.321 "num_base_bdevs_discovered": 1, 00:26:37.321 "num_base_bdevs_operational": 1, 00:26:37.321 "base_bdevs_list": [ 00:26:37.321 { 00:26:37.321 "name": null, 00:26:37.321 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:37.321 "is_configured": false, 00:26:37.321 "data_offset": 0, 00:26:37.321 "data_size": 63488 00:26:37.321 }, 00:26:37.321 { 00:26:37.321 "name": "BaseBdev2", 00:26:37.321 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:37.321 "is_configured": true, 00:26:37.321 "data_offset": 2048, 00:26:37.321 "data_size": 63488 00:26:37.321 } 00:26:37.321 ] 00:26:37.321 }' 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:37.321 [2024-10-01 20:25:32.515410] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:26:37.321 [2024-10-01 20:25:32.515484] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:37.321 [2024-10-01 20:25:32.515542] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:26:37.321 [2024-10-01 20:25:32.515562] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:37.321 [2024-10-01 20:25:32.516190] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:37.321 [2024-10-01 20:25:32.516228] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:26:37.321 [2024-10-01 20:25:32.516345] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:26:37.321 [2024-10-01 20:25:32.516369] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:26:37.321 [2024-10-01 20:25:32.516384] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:26:37.321 [2024-10-01 20:25:32.516402] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:26:37.321 BaseBdev1 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:37.321 20:25:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:38.698 "name": "raid_bdev1", 00:26:38.698 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:38.698 "strip_size_kb": 0, 00:26:38.698 "state": "online", 00:26:38.698 "raid_level": "raid1", 00:26:38.698 "superblock": true, 00:26:38.698 "num_base_bdevs": 2, 00:26:38.698 "num_base_bdevs_discovered": 1, 00:26:38.698 "num_base_bdevs_operational": 1, 00:26:38.698 "base_bdevs_list": [ 00:26:38.698 { 00:26:38.698 "name": null, 00:26:38.698 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:38.698 "is_configured": false, 00:26:38.698 "data_offset": 0, 00:26:38.698 "data_size": 63488 00:26:38.698 }, 00:26:38.698 { 00:26:38.698 "name": "BaseBdev2", 00:26:38.698 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:38.698 "is_configured": true, 00:26:38.698 "data_offset": 2048, 00:26:38.698 "data_size": 63488 00:26:38.698 } 00:26:38.698 ] 00:26:38.698 }' 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:38.698 20:25:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:38.957 20:25:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:26:38.957 20:25:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:38.957 20:25:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:26:38.957 20:25:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:26:38.957 20:25:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:38.957 20:25:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:38.957 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:38.957 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:38.957 20:25:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:38.957 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:38.957 20:25:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:38.957 "name": "raid_bdev1", 00:26:38.957 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:38.957 "strip_size_kb": 0, 00:26:38.957 "state": "online", 00:26:38.957 "raid_level": "raid1", 00:26:38.957 "superblock": true, 00:26:38.957 "num_base_bdevs": 2, 00:26:38.957 "num_base_bdevs_discovered": 1, 00:26:38.957 "num_base_bdevs_operational": 1, 00:26:38.957 "base_bdevs_list": [ 00:26:38.957 { 00:26:38.957 "name": null, 00:26:38.957 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:38.957 "is_configured": false, 00:26:38.957 "data_offset": 0, 00:26:38.957 "data_size": 63488 00:26:38.957 }, 00:26:38.957 { 00:26:38.957 "name": "BaseBdev2", 00:26:38.957 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:38.957 "is_configured": true, 00:26:38.957 "data_offset": 2048, 00:26:38.957 "data_size": 63488 00:26:38.957 } 00:26:38.957 ] 00:26:38.957 }' 00:26:38.957 20:25:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:38.957 20:25:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:26:38.957 20:25:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:39.217 [2024-10-01 20:25:34.280055] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:39.217 [2024-10-01 20:25:34.280315] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:26:39.217 [2024-10-01 20:25:34.280342] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:26:39.217 request: 00:26:39.217 { 00:26:39.217 "base_bdev": "BaseBdev1", 00:26:39.217 "raid_bdev": "raid_bdev1", 00:26:39.217 "method": "bdev_raid_add_base_bdev", 00:26:39.217 "req_id": 1 00:26:39.217 } 00:26:39.217 Got JSON-RPC error response 00:26:39.217 response: 00:26:39.217 { 00:26:39.217 "code": -22, 00:26:39.217 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:26:39.217 } 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:26:39.217 20:25:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:40.193 "name": "raid_bdev1", 00:26:40.193 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:40.193 "strip_size_kb": 0, 00:26:40.193 "state": "online", 00:26:40.193 "raid_level": "raid1", 00:26:40.193 "superblock": true, 00:26:40.193 "num_base_bdevs": 2, 00:26:40.193 "num_base_bdevs_discovered": 1, 00:26:40.193 "num_base_bdevs_operational": 1, 00:26:40.193 "base_bdevs_list": [ 00:26:40.193 { 00:26:40.193 "name": null, 00:26:40.193 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:40.193 "is_configured": false, 00:26:40.193 "data_offset": 0, 00:26:40.193 "data_size": 63488 00:26:40.193 }, 00:26:40.193 { 00:26:40.193 "name": "BaseBdev2", 00:26:40.193 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:40.193 "is_configured": true, 00:26:40.193 "data_offset": 2048, 00:26:40.193 "data_size": 63488 00:26:40.193 } 00:26:40.193 ] 00:26:40.193 }' 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:40.193 20:25:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:40.759 "name": "raid_bdev1", 00:26:40.759 "uuid": "955524fa-51ab-49be-90d7-6b54c7cf8d61", 00:26:40.759 "strip_size_kb": 0, 00:26:40.759 "state": "online", 00:26:40.759 "raid_level": "raid1", 00:26:40.759 "superblock": true, 00:26:40.759 "num_base_bdevs": 2, 00:26:40.759 "num_base_bdevs_discovered": 1, 00:26:40.759 "num_base_bdevs_operational": 1, 00:26:40.759 "base_bdevs_list": [ 00:26:40.759 { 00:26:40.759 "name": null, 00:26:40.759 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:40.759 "is_configured": false, 00:26:40.759 "data_offset": 0, 00:26:40.759 "data_size": 63488 00:26:40.759 }, 00:26:40.759 { 00:26:40.759 "name": "BaseBdev2", 00:26:40.759 "uuid": "63b879c8-f456-52ff-a1e1-20987fcc87c9", 00:26:40.759 "is_configured": true, 00:26:40.759 "data_offset": 2048, 00:26:40.759 "data_size": 63488 00:26:40.759 } 00:26:40.759 ] 00:26:40.759 }' 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 76682 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@950 -- # '[' -z 76682 ']' 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # kill -0 76682 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@955 -- # uname 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:26:40.759 20:25:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 76682 00:26:41.016 20:25:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:26:41.016 20:25:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:26:41.016 killing process with pid 76682 00:26:41.016 20:25:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 76682' 00:26:41.017 Received shutdown signal, test time was about 60.000000 seconds 00:26:41.017 00:26:41.017 Latency(us) 00:26:41.017 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:41.017 =================================================================================================================== 00:26:41.017 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:26:41.017 20:25:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@969 -- # kill 76682 00:26:41.017 [2024-10-01 20:25:36.019914] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:41.017 [2024-10-01 20:25:36.020090] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:41.017 [2024-10-01 20:25:36.020175] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:41.017 [2024-10-01 20:25:36.020199] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:26:41.017 20:25:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@974 -- # wait 76682 00:26:41.275 [2024-10-01 20:25:36.290699] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:26:43.176 00:26:43.176 real 0m28.101s 00:26:43.176 user 0m34.438s 00:26:43.176 sys 0m4.375s 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:26:43.176 ************************************ 00:26:43.176 END TEST raid_rebuild_test_sb 00:26:43.176 ************************************ 00:26:43.176 20:25:38 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 2 false true true 00:26:43.176 20:25:38 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:26:43.176 20:25:38 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:26:43.176 20:25:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:43.176 ************************************ 00:26:43.176 START TEST raid_rebuild_test_io 00:26:43.176 ************************************ 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 false true true 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=77451 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 77451 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@831 -- # '[' -z 77451 ']' 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:26:43.176 20:25:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@836 -- # local max_retries=100 00:26:43.176 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:43.177 20:25:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:43.177 20:25:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # xtrace_disable 00:26:43.177 20:25:38 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:43.177 [2024-10-01 20:25:38.326233] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:26:43.177 I/O size of 3145728 is greater than zero copy threshold (65536). 00:26:43.177 Zero copy mechanism will not be used. 00:26:43.177 [2024-10-01 20:25:38.326406] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77451 ] 00:26:43.435 [2024-10-01 20:25:38.498222] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:43.692 [2024-10-01 20:25:38.746216] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:26:43.953 [2024-10-01 20:25:38.961141] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:43.953 [2024-10-01 20:25:38.961248] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:44.213 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:26:44.213 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # return 0 00:26:44.213 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:26:44.213 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:26:44.213 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:44.213 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:44.213 BaseBdev1_malloc 00:26:44.213 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:44.213 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:26:44.213 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:44.213 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:44.213 [2024-10-01 20:25:39.420655] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:26:44.213 [2024-10-01 20:25:39.420783] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:44.213 [2024-10-01 20:25:39.420824] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:26:44.213 [2024-10-01 20:25:39.420851] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:44.213 [2024-10-01 20:25:39.423993] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:44.213 [2024-10-01 20:25:39.424050] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:26:44.213 BaseBdev1 00:26:44.213 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:44.213 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:26:44.213 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:26:44.213 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:44.213 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:44.519 BaseBdev2_malloc 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:44.519 [2024-10-01 20:25:39.479657] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:26:44.519 [2024-10-01 20:25:39.479758] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:44.519 [2024-10-01 20:25:39.479802] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:26:44.519 [2024-10-01 20:25:39.479824] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:44.519 [2024-10-01 20:25:39.482830] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:44.519 [2024-10-01 20:25:39.482887] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:26:44.519 BaseBdev2 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:44.519 spare_malloc 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:44.519 spare_delay 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:44.519 [2024-10-01 20:25:39.540085] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:26:44.519 [2024-10-01 20:25:39.540176] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:44.519 [2024-10-01 20:25:39.540210] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:26:44.519 [2024-10-01 20:25:39.540230] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:44.519 [2024-10-01 20:25:39.543183] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:44.519 [2024-10-01 20:25:39.543238] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:26:44.519 spare 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:44.519 [2024-10-01 20:25:39.548224] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:26:44.519 [2024-10-01 20:25:39.550917] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:26:44.519 [2024-10-01 20:25:39.551065] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:26:44.519 [2024-10-01 20:25:39.551087] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:26:44.519 [2024-10-01 20:25:39.551455] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:26:44.519 [2024-10-01 20:25:39.551672] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:26:44.519 [2024-10-01 20:25:39.551689] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:26:44.519 [2024-10-01 20:25:39.552081] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:44.519 "name": "raid_bdev1", 00:26:44.519 "uuid": "c81fa02b-dab0-4187-8421-45ab65c019b0", 00:26:44.519 "strip_size_kb": 0, 00:26:44.519 "state": "online", 00:26:44.519 "raid_level": "raid1", 00:26:44.519 "superblock": false, 00:26:44.519 "num_base_bdevs": 2, 00:26:44.519 "num_base_bdevs_discovered": 2, 00:26:44.519 "num_base_bdevs_operational": 2, 00:26:44.519 "base_bdevs_list": [ 00:26:44.519 { 00:26:44.519 "name": "BaseBdev1", 00:26:44.519 "uuid": "e017a96f-ede1-5aff-a2d5-ad16ece354aa", 00:26:44.519 "is_configured": true, 00:26:44.519 "data_offset": 0, 00:26:44.519 "data_size": 65536 00:26:44.519 }, 00:26:44.519 { 00:26:44.519 "name": "BaseBdev2", 00:26:44.519 "uuid": "35991d7f-c7cb-5bdf-a9d8-86e42dcde592", 00:26:44.519 "is_configured": true, 00:26:44.519 "data_offset": 0, 00:26:44.519 "data_size": 65536 00:26:44.519 } 00:26:44.519 ] 00:26:44.519 }' 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:44.519 20:25:39 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:45.109 [2024-10-01 20:25:40.092794] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:45.109 [2024-10-01 20:25:40.192428] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:45.109 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:45.109 "name": "raid_bdev1", 00:26:45.109 "uuid": "c81fa02b-dab0-4187-8421-45ab65c019b0", 00:26:45.109 "strip_size_kb": 0, 00:26:45.109 "state": "online", 00:26:45.109 "raid_level": "raid1", 00:26:45.109 "superblock": false, 00:26:45.110 "num_base_bdevs": 2, 00:26:45.110 "num_base_bdevs_discovered": 1, 00:26:45.110 "num_base_bdevs_operational": 1, 00:26:45.110 "base_bdevs_list": [ 00:26:45.110 { 00:26:45.110 "name": null, 00:26:45.110 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:45.110 "is_configured": false, 00:26:45.110 "data_offset": 0, 00:26:45.110 "data_size": 65536 00:26:45.110 }, 00:26:45.110 { 00:26:45.110 "name": "BaseBdev2", 00:26:45.110 "uuid": "35991d7f-c7cb-5bdf-a9d8-86e42dcde592", 00:26:45.110 "is_configured": true, 00:26:45.110 "data_offset": 0, 00:26:45.110 "data_size": 65536 00:26:45.110 } 00:26:45.110 ] 00:26:45.110 }' 00:26:45.110 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:45.110 20:25:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:45.110 [2024-10-01 20:25:40.300781] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:26:45.110 I/O size of 3145728 is greater than zero copy threshold (65536). 00:26:45.110 Zero copy mechanism will not be used. 00:26:45.110 Running I/O for 60 seconds... 00:26:45.676 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:26:45.676 20:25:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:45.676 20:25:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:45.676 [2024-10-01 20:25:40.686933] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:26:45.676 20:25:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:45.676 20:25:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:26:45.676 [2024-10-01 20:25:40.750245] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:26:45.676 [2024-10-01 20:25:40.752997] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:26:45.676 [2024-10-01 20:25:40.878511] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:26:45.676 [2024-10-01 20:25:40.879406] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:26:46.242 178.00 IOPS, 534.00 MiB/s [2024-10-01 20:25:41.396994] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:26:46.242 [2024-10-01 20:25:41.397723] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:26:46.501 [2024-10-01 20:25:41.609513] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:26:46.501 [2024-10-01 20:25:41.610003] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:26:46.501 20:25:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:46.501 20:25:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:46.501 20:25:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:46.501 20:25:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:46.501 20:25:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:46.501 20:25:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:46.501 20:25:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:46.501 20:25:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:46.501 20:25:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:46.760 20:25:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:46.760 20:25:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:46.760 "name": "raid_bdev1", 00:26:46.760 "uuid": "c81fa02b-dab0-4187-8421-45ab65c019b0", 00:26:46.760 "strip_size_kb": 0, 00:26:46.760 "state": "online", 00:26:46.760 "raid_level": "raid1", 00:26:46.760 "superblock": false, 00:26:46.760 "num_base_bdevs": 2, 00:26:46.760 "num_base_bdevs_discovered": 2, 00:26:46.760 "num_base_bdevs_operational": 2, 00:26:46.760 "process": { 00:26:46.760 "type": "rebuild", 00:26:46.760 "target": "spare", 00:26:46.760 "progress": { 00:26:46.760 "blocks": 12288, 00:26:46.760 "percent": 18 00:26:46.760 } 00:26:46.760 }, 00:26:46.760 "base_bdevs_list": [ 00:26:46.760 { 00:26:46.760 "name": "spare", 00:26:46.760 "uuid": "7434ef83-717a-55fd-83f6-e839fa3e83a2", 00:26:46.760 "is_configured": true, 00:26:46.760 "data_offset": 0, 00:26:46.760 "data_size": 65536 00:26:46.760 }, 00:26:46.760 { 00:26:46.760 "name": "BaseBdev2", 00:26:46.760 "uuid": "35991d7f-c7cb-5bdf-a9d8-86e42dcde592", 00:26:46.760 "is_configured": true, 00:26:46.760 "data_offset": 0, 00:26:46.760 "data_size": 65536 00:26:46.760 } 00:26:46.760 ] 00:26:46.760 }' 00:26:46.760 20:25:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:46.760 20:25:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:46.760 20:25:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:46.760 [2024-10-01 20:25:41.868404] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:26:46.760 20:25:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:26:46.760 20:25:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:26:46.760 20:25:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:46.760 20:25:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:46.760 [2024-10-01 20:25:41.901795] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:47.018 [2024-10-01 20:25:42.086939] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:26:47.018 [2024-10-01 20:25:42.090258] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:47.018 [2024-10-01 20:25:42.090473] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:26:47.018 [2024-10-01 20:25:42.090636] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:26:47.018 [2024-10-01 20:25:42.137038] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006080 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:47.018 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:47.018 "name": "raid_bdev1", 00:26:47.018 "uuid": "c81fa02b-dab0-4187-8421-45ab65c019b0", 00:26:47.018 "strip_size_kb": 0, 00:26:47.018 "state": "online", 00:26:47.018 "raid_level": "raid1", 00:26:47.018 "superblock": false, 00:26:47.018 "num_base_bdevs": 2, 00:26:47.018 "num_base_bdevs_discovered": 1, 00:26:47.019 "num_base_bdevs_operational": 1, 00:26:47.019 "base_bdevs_list": [ 00:26:47.019 { 00:26:47.019 "name": null, 00:26:47.019 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:47.019 "is_configured": false, 00:26:47.019 "data_offset": 0, 00:26:47.019 "data_size": 65536 00:26:47.019 }, 00:26:47.019 { 00:26:47.019 "name": "BaseBdev2", 00:26:47.019 "uuid": "35991d7f-c7cb-5bdf-a9d8-86e42dcde592", 00:26:47.019 "is_configured": true, 00:26:47.019 "data_offset": 0, 00:26:47.019 "data_size": 65536 00:26:47.019 } 00:26:47.019 ] 00:26:47.019 }' 00:26:47.019 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:47.019 20:25:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:47.536 136.00 IOPS, 408.00 MiB/s 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:26:47.536 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:47.536 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:26:47.536 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:26:47.536 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:47.536 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:47.536 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:47.536 20:25:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:47.536 20:25:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:47.536 20:25:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:47.536 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:47.536 "name": "raid_bdev1", 00:26:47.536 "uuid": "c81fa02b-dab0-4187-8421-45ab65c019b0", 00:26:47.536 "strip_size_kb": 0, 00:26:47.536 "state": "online", 00:26:47.536 "raid_level": "raid1", 00:26:47.536 "superblock": false, 00:26:47.536 "num_base_bdevs": 2, 00:26:47.536 "num_base_bdevs_discovered": 1, 00:26:47.536 "num_base_bdevs_operational": 1, 00:26:47.536 "base_bdevs_list": [ 00:26:47.536 { 00:26:47.536 "name": null, 00:26:47.536 "uuid": "00000000-0000-0000-0000-000000000000", 00:26:47.536 "is_configured": false, 00:26:47.536 "data_offset": 0, 00:26:47.536 "data_size": 65536 00:26:47.536 }, 00:26:47.536 { 00:26:47.536 "name": "BaseBdev2", 00:26:47.536 "uuid": "35991d7f-c7cb-5bdf-a9d8-86e42dcde592", 00:26:47.536 "is_configured": true, 00:26:47.536 "data_offset": 0, 00:26:47.536 "data_size": 65536 00:26:47.536 } 00:26:47.536 ] 00:26:47.536 }' 00:26:47.536 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:47.536 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:26:47.795 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:47.795 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:26:47.795 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:26:47.795 20:25:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:47.795 20:25:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:47.795 [2024-10-01 20:25:42.853088] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:26:47.795 20:25:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:47.795 20:25:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:26:47.795 [2024-10-01 20:25:42.922269] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:26:47.795 [2024-10-01 20:25:42.925346] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:26:47.795 [2024-10-01 20:25:43.044382] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:26:47.795 [2024-10-01 20:25:43.045163] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:26:48.053 [2024-10-01 20:25:43.275375] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:26:48.053 [2024-10-01 20:25:43.275836] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:26:48.572 148.00 IOPS, 444.00 MiB/s [2024-10-01 20:25:43.743969] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:26:48.572 [2024-10-01 20:25:43.744391] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:26:48.833 20:25:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:48.833 20:25:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:48.833 20:25:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:48.833 20:25:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:48.833 20:25:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:48.833 20:25:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:48.833 20:25:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:48.833 20:25:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:48.833 20:25:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:48.833 20:25:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:48.833 20:25:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:48.833 "name": "raid_bdev1", 00:26:48.833 "uuid": "c81fa02b-dab0-4187-8421-45ab65c019b0", 00:26:48.833 "strip_size_kb": 0, 00:26:48.833 "state": "online", 00:26:48.833 "raid_level": "raid1", 00:26:48.833 "superblock": false, 00:26:48.833 "num_base_bdevs": 2, 00:26:48.833 "num_base_bdevs_discovered": 2, 00:26:48.833 "num_base_bdevs_operational": 2, 00:26:48.833 "process": { 00:26:48.833 "type": "rebuild", 00:26:48.833 "target": "spare", 00:26:48.833 "progress": { 00:26:48.833 "blocks": 10240, 00:26:48.833 "percent": 15 00:26:48.833 } 00:26:48.833 }, 00:26:48.833 "base_bdevs_list": [ 00:26:48.833 { 00:26:48.833 "name": "spare", 00:26:48.833 "uuid": "7434ef83-717a-55fd-83f6-e839fa3e83a2", 00:26:48.833 "is_configured": true, 00:26:48.833 "data_offset": 0, 00:26:48.833 "data_size": 65536 00:26:48.833 }, 00:26:48.833 { 00:26:48.833 "name": "BaseBdev2", 00:26:48.833 "uuid": "35991d7f-c7cb-5bdf-a9d8-86e42dcde592", 00:26:48.833 "is_configured": true, 00:26:48.833 "data_offset": 0, 00:26:48.833 "data_size": 65536 00:26:48.833 } 00:26:48.833 ] 00:26:48.833 }' 00:26:48.833 20:25:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=483 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:48.833 20:25:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:49.092 20:25:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:49.092 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:49.092 "name": "raid_bdev1", 00:26:49.092 "uuid": "c81fa02b-dab0-4187-8421-45ab65c019b0", 00:26:49.092 "strip_size_kb": 0, 00:26:49.092 "state": "online", 00:26:49.092 "raid_level": "raid1", 00:26:49.092 "superblock": false, 00:26:49.092 "num_base_bdevs": 2, 00:26:49.092 "num_base_bdevs_discovered": 2, 00:26:49.092 "num_base_bdevs_operational": 2, 00:26:49.092 "process": { 00:26:49.092 "type": "rebuild", 00:26:49.092 "target": "spare", 00:26:49.092 "progress": { 00:26:49.092 "blocks": 12288, 00:26:49.092 "percent": 18 00:26:49.092 } 00:26:49.092 }, 00:26:49.092 "base_bdevs_list": [ 00:26:49.092 { 00:26:49.092 "name": "spare", 00:26:49.092 "uuid": "7434ef83-717a-55fd-83f6-e839fa3e83a2", 00:26:49.092 "is_configured": true, 00:26:49.092 "data_offset": 0, 00:26:49.092 "data_size": 65536 00:26:49.092 }, 00:26:49.092 { 00:26:49.092 "name": "BaseBdev2", 00:26:49.092 "uuid": "35991d7f-c7cb-5bdf-a9d8-86e42dcde592", 00:26:49.092 "is_configured": true, 00:26:49.092 "data_offset": 0, 00:26:49.092 "data_size": 65536 00:26:49.092 } 00:26:49.092 ] 00:26:49.092 }' 00:26:49.092 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:49.092 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:49.092 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:49.092 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:26:49.092 20:25:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:26:49.351 131.75 IOPS, 395.25 MiB/s [2024-10-01 20:25:44.440618] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:26:49.351 [2024-10-01 20:25:44.542566] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:26:49.609 [2024-10-01 20:25:44.787524] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:26:49.609 [2024-10-01 20:25:44.788175] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:26:49.867 [2024-10-01 20:25:45.017721] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:26:50.126 20:25:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:26:50.126 20:25:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:50.126 20:25:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:50.126 20:25:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:50.126 20:25:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:50.126 20:25:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:50.126 20:25:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:50.126 20:25:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:50.126 20:25:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:50.126 20:25:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:50.126 20:25:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:50.126 20:25:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:50.126 "name": "raid_bdev1", 00:26:50.126 "uuid": "c81fa02b-dab0-4187-8421-45ab65c019b0", 00:26:50.126 "strip_size_kb": 0, 00:26:50.126 "state": "online", 00:26:50.126 "raid_level": "raid1", 00:26:50.126 "superblock": false, 00:26:50.126 "num_base_bdevs": 2, 00:26:50.126 "num_base_bdevs_discovered": 2, 00:26:50.126 "num_base_bdevs_operational": 2, 00:26:50.126 "process": { 00:26:50.126 "type": "rebuild", 00:26:50.126 "target": "spare", 00:26:50.126 "progress": { 00:26:50.126 "blocks": 28672, 00:26:50.126 "percent": 43 00:26:50.126 } 00:26:50.126 }, 00:26:50.126 "base_bdevs_list": [ 00:26:50.126 { 00:26:50.126 "name": "spare", 00:26:50.126 "uuid": "7434ef83-717a-55fd-83f6-e839fa3e83a2", 00:26:50.126 "is_configured": true, 00:26:50.126 "data_offset": 0, 00:26:50.126 "data_size": 65536 00:26:50.126 }, 00:26:50.126 { 00:26:50.126 "name": "BaseBdev2", 00:26:50.126 "uuid": "35991d7f-c7cb-5bdf-a9d8-86e42dcde592", 00:26:50.126 "is_configured": true, 00:26:50.126 "data_offset": 0, 00:26:50.126 "data_size": 65536 00:26:50.126 } 00:26:50.126 ] 00:26:50.126 }' 00:26:50.126 20:25:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:50.126 112.80 IOPS, 338.40 MiB/s 20:25:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:50.126 20:25:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:50.126 [2024-10-01 20:25:45.368070] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:26:50.385 20:25:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:26:50.385 20:25:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:26:50.385 [2024-10-01 20:25:45.488025] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:26:50.644 [2024-10-01 20:25:45.730176] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:26:50.902 [2024-10-01 20:25:45.934915] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:26:50.902 [2024-10-01 20:25:45.935353] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:26:51.161 [2024-10-01 20:25:46.304825] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:26:51.161 101.67 IOPS, 305.00 MiB/s 20:25:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:26:51.161 20:25:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:51.161 20:25:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:51.161 20:25:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:51.161 20:25:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:51.161 20:25:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:51.161 20:25:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:51.161 20:25:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:51.161 20:25:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:51.161 20:25:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:51.420 20:25:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:51.420 20:25:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:51.420 "name": "raid_bdev1", 00:26:51.420 "uuid": "c81fa02b-dab0-4187-8421-45ab65c019b0", 00:26:51.420 "strip_size_kb": 0, 00:26:51.420 "state": "online", 00:26:51.420 "raid_level": "raid1", 00:26:51.420 "superblock": false, 00:26:51.420 "num_base_bdevs": 2, 00:26:51.420 "num_base_bdevs_discovered": 2, 00:26:51.420 "num_base_bdevs_operational": 2, 00:26:51.420 "process": { 00:26:51.420 "type": "rebuild", 00:26:51.420 "target": "spare", 00:26:51.420 "progress": { 00:26:51.420 "blocks": 47104, 00:26:51.420 "percent": 71 00:26:51.420 } 00:26:51.420 }, 00:26:51.420 "base_bdevs_list": [ 00:26:51.420 { 00:26:51.420 "name": "spare", 00:26:51.420 "uuid": "7434ef83-717a-55fd-83f6-e839fa3e83a2", 00:26:51.420 "is_configured": true, 00:26:51.420 "data_offset": 0, 00:26:51.420 "data_size": 65536 00:26:51.420 }, 00:26:51.421 { 00:26:51.421 "name": "BaseBdev2", 00:26:51.421 "uuid": "35991d7f-c7cb-5bdf-a9d8-86e42dcde592", 00:26:51.421 "is_configured": true, 00:26:51.421 "data_offset": 0, 00:26:51.421 "data_size": 65536 00:26:51.421 } 00:26:51.421 ] 00:26:51.421 }' 00:26:51.421 20:25:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:51.421 20:25:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:51.421 20:25:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:51.421 20:25:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:26:51.421 20:25:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:26:51.679 [2024-10-01 20:25:46.763290] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:26:51.939 [2024-10-01 20:25:47.092214] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:26:52.457 92.00 IOPS, 276.00 MiB/s [2024-10-01 20:25:47.544268] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:26:52.457 20:25:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:26:52.457 20:25:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:52.457 20:25:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:52.457 20:25:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:52.457 20:25:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:52.457 20:25:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:52.457 20:25:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:52.457 20:25:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:52.457 20:25:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:52.457 20:25:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:52.457 20:25:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:52.457 20:25:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:52.457 "name": "raid_bdev1", 00:26:52.457 "uuid": "c81fa02b-dab0-4187-8421-45ab65c019b0", 00:26:52.457 "strip_size_kb": 0, 00:26:52.457 "state": "online", 00:26:52.457 "raid_level": "raid1", 00:26:52.457 "superblock": false, 00:26:52.457 "num_base_bdevs": 2, 00:26:52.457 "num_base_bdevs_discovered": 2, 00:26:52.457 "num_base_bdevs_operational": 2, 00:26:52.457 "process": { 00:26:52.457 "type": "rebuild", 00:26:52.457 "target": "spare", 00:26:52.457 "progress": { 00:26:52.457 "blocks": 65536, 00:26:52.457 "percent": 100 00:26:52.457 } 00:26:52.457 }, 00:26:52.457 "base_bdevs_list": [ 00:26:52.457 { 00:26:52.457 "name": "spare", 00:26:52.457 "uuid": "7434ef83-717a-55fd-83f6-e839fa3e83a2", 00:26:52.457 "is_configured": true, 00:26:52.457 "data_offset": 0, 00:26:52.457 "data_size": 65536 00:26:52.457 }, 00:26:52.457 { 00:26:52.457 "name": "BaseBdev2", 00:26:52.457 "uuid": "35991d7f-c7cb-5bdf-a9d8-86e42dcde592", 00:26:52.457 "is_configured": true, 00:26:52.457 "data_offset": 0, 00:26:52.457 "data_size": 65536 00:26:52.457 } 00:26:52.457 ] 00:26:52.457 }' 00:26:52.457 20:25:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:52.457 [2024-10-01 20:25:47.644262] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:26:52.457 [2024-10-01 20:25:47.647043] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:52.457 20:25:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:26:52.457 20:25:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:52.716 20:25:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:26:52.716 20:25:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:26:53.541 85.62 IOPS, 256.88 MiB/s 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:26:53.541 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:26:53.541 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:53.541 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:26:53.541 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:26:53.541 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:53.542 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:53.542 20:25:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:53.542 20:25:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:53.542 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:53.542 20:25:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:53.800 "name": "raid_bdev1", 00:26:53.800 "uuid": "c81fa02b-dab0-4187-8421-45ab65c019b0", 00:26:53.800 "strip_size_kb": 0, 00:26:53.800 "state": "online", 00:26:53.800 "raid_level": "raid1", 00:26:53.800 "superblock": false, 00:26:53.800 "num_base_bdevs": 2, 00:26:53.800 "num_base_bdevs_discovered": 2, 00:26:53.800 "num_base_bdevs_operational": 2, 00:26:53.800 "base_bdevs_list": [ 00:26:53.800 { 00:26:53.800 "name": "spare", 00:26:53.800 "uuid": "7434ef83-717a-55fd-83f6-e839fa3e83a2", 00:26:53.800 "is_configured": true, 00:26:53.800 "data_offset": 0, 00:26:53.800 "data_size": 65536 00:26:53.800 }, 00:26:53.800 { 00:26:53.800 "name": "BaseBdev2", 00:26:53.800 "uuid": "35991d7f-c7cb-5bdf-a9d8-86e42dcde592", 00:26:53.800 "is_configured": true, 00:26:53.800 "data_offset": 0, 00:26:53.800 "data_size": 65536 00:26:53.800 } 00:26:53.800 ] 00:26:53.800 }' 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:26:53.800 "name": "raid_bdev1", 00:26:53.800 "uuid": "c81fa02b-dab0-4187-8421-45ab65c019b0", 00:26:53.800 "strip_size_kb": 0, 00:26:53.800 "state": "online", 00:26:53.800 "raid_level": "raid1", 00:26:53.800 "superblock": false, 00:26:53.800 "num_base_bdevs": 2, 00:26:53.800 "num_base_bdevs_discovered": 2, 00:26:53.800 "num_base_bdevs_operational": 2, 00:26:53.800 "base_bdevs_list": [ 00:26:53.800 { 00:26:53.800 "name": "spare", 00:26:53.800 "uuid": "7434ef83-717a-55fd-83f6-e839fa3e83a2", 00:26:53.800 "is_configured": true, 00:26:53.800 "data_offset": 0, 00:26:53.800 "data_size": 65536 00:26:53.800 }, 00:26:53.800 { 00:26:53.800 "name": "BaseBdev2", 00:26:53.800 "uuid": "35991d7f-c7cb-5bdf-a9d8-86e42dcde592", 00:26:53.800 "is_configured": true, 00:26:53.800 "data_offset": 0, 00:26:53.800 "data_size": 65536 00:26:53.800 } 00:26:53.800 ] 00:26:53.800 }' 00:26:53.800 20:25:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:26:53.800 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:26:53.800 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:26:54.059 "name": "raid_bdev1", 00:26:54.059 "uuid": "c81fa02b-dab0-4187-8421-45ab65c019b0", 00:26:54.059 "strip_size_kb": 0, 00:26:54.059 "state": "online", 00:26:54.059 "raid_level": "raid1", 00:26:54.059 "superblock": false, 00:26:54.059 "num_base_bdevs": 2, 00:26:54.059 "num_base_bdevs_discovered": 2, 00:26:54.059 "num_base_bdevs_operational": 2, 00:26:54.059 "base_bdevs_list": [ 00:26:54.059 { 00:26:54.059 "name": "spare", 00:26:54.059 "uuid": "7434ef83-717a-55fd-83f6-e839fa3e83a2", 00:26:54.059 "is_configured": true, 00:26:54.059 "data_offset": 0, 00:26:54.059 "data_size": 65536 00:26:54.059 }, 00:26:54.059 { 00:26:54.059 "name": "BaseBdev2", 00:26:54.059 "uuid": "35991d7f-c7cb-5bdf-a9d8-86e42dcde592", 00:26:54.059 "is_configured": true, 00:26:54.059 "data_offset": 0, 00:26:54.059 "data_size": 65536 00:26:54.059 } 00:26:54.059 ] 00:26:54.059 }' 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:26:54.059 20:25:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:54.577 79.67 IOPS, 239.00 MiB/s 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:54.577 [2024-10-01 20:25:49.592822] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:26:54.577 [2024-10-01 20:25:49.592860] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:26:54.577 00:26:54.577 Latency(us) 00:26:54.577 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:54.577 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:26:54.577 raid_bdev1 : 9.33 77.82 233.47 0.00 0.00 17557.19 316.51 118203.11 00:26:54.577 =================================================================================================================== 00:26:54.577 Total : 77.82 233.47 0.00 0.00 17557.19 316.51 118203.11 00:26:54.577 [2024-10-01 20:25:49.654438] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:26:54.577 [2024-10-01 20:25:49.654513] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:26:54.577 [2024-10-01 20:25:49.654634] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:26:54.577 [2024-10-01 20:25:49.654652] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:26:54.577 { 00:26:54.577 "results": [ 00:26:54.577 { 00:26:54.577 "job": "raid_bdev1", 00:26:54.577 "core_mask": "0x1", 00:26:54.577 "workload": "randrw", 00:26:54.577 "percentage": 50, 00:26:54.577 "status": "finished", 00:26:54.577 "queue_depth": 2, 00:26:54.577 "io_size": 3145728, 00:26:54.577 "runtime": 9.32867, 00:26:54.577 "iops": 77.82459879061003, 00:26:54.577 "mibps": 233.47379637183008, 00:26:54.577 "io_failed": 0, 00:26:54.577 "io_timeout": 0, 00:26:54.577 "avg_latency_us": 17557.187798647636, 00:26:54.577 "min_latency_us": 316.5090909090909, 00:26:54.577 "max_latency_us": 118203.11272727273 00:26:54.577 } 00:26:54.577 ], 00:26:54.577 "core_count": 1 00:26:54.577 } 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:54.577 20:25:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:26:54.837 /dev/nbd0 00:26:54.837 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:26:54.837 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:26:54.837 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:26:54.837 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:26:54.837 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:26:54.837 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:26:54.837 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:26:55.096 1+0 records in 00:26:55.096 1+0 records out 00:26:55.096 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000377412 s, 10.9 MB/s 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:55.096 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:26:55.358 /dev/nbd1 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:26:55.358 1+0 records in 00:26:55.358 1+0 records out 00:26:55.358 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000380595 s, 10.8 MB/s 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:26:55.358 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:26:55.618 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:26:55.618 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:26:55.618 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:26:55.618 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:26:55.618 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:26:55.618 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:26:55.618 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:26:55.877 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:26:55.877 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:26:55.877 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:26:55.877 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:26:55.877 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:26:55.877 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:26:55.877 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:26:55.877 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:26:55.877 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:26:55.877 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:26:55.877 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:26:55.877 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:26:55.877 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:26:55.877 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:26:55.877 20:25:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 77451 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@950 -- # '[' -z 77451 ']' 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # kill -0 77451 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@955 -- # uname 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 77451 00:26:56.136 killing process with pid 77451 00:26:56.136 Received shutdown signal, test time was about 11.006152 seconds 00:26:56.136 00:26:56.136 Latency(us) 00:26:56.136 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:56.136 =================================================================================================================== 00:26:56.136 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@968 -- # echo 'killing process with pid 77451' 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@969 -- # kill 77451 00:26:56.136 [2024-10-01 20:25:51.310144] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:26:56.136 20:25:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@974 -- # wait 77451 00:26:56.395 [2024-10-01 20:25:51.542594] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:26:58.926 00:26:58.926 real 0m15.349s 00:26:58.926 user 0m19.464s 00:26:58.926 sys 0m1.671s 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:26:58.926 ************************************ 00:26:58.926 END TEST raid_rebuild_test_io 00:26:58.926 ************************************ 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:26:58.926 20:25:53 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 2 true true true 00:26:58.926 20:25:53 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:26:58.926 20:25:53 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:26:58.926 20:25:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:26:58.926 ************************************ 00:26:58.926 START TEST raid_rebuild_test_sb_io 00:26:58.926 ************************************ 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true true true 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:26:58.926 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=77867 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 77867 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@831 -- # '[' -z 77867 ']' 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@836 -- # local max_retries=100 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:58.926 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # xtrace_disable 00:26:58.927 20:25:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:26:58.927 [2024-10-01 20:25:53.742147] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:26:58.927 I/O size of 3145728 is greater than zero copy threshold (65536). 00:26:58.927 Zero copy mechanism will not be used. 00:26:58.927 [2024-10-01 20:25:53.742580] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77867 ] 00:26:58.927 [2024-10-01 20:25:53.921173] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:58.927 [2024-10-01 20:25:54.176707] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:26:59.185 [2024-10-01 20:25:54.380883] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:59.185 [2024-10-01 20:25:54.380979] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # return 0 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:26:59.753 BaseBdev1_malloc 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:26:59.753 [2024-10-01 20:25:54.918661] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:26:59.753 [2024-10-01 20:25:54.918909] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:59.753 [2024-10-01 20:25:54.918958] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:26:59.753 [2024-10-01 20:25:54.918987] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:59.753 [2024-10-01 20:25:54.922454] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:59.753 BaseBdev1 00:26:59.753 [2024-10-01 20:25:54.922647] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:26:59.753 BaseBdev2_malloc 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:26:59.753 [2024-10-01 20:25:54.971357] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:26:59.753 [2024-10-01 20:25:54.971443] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:26:59.753 [2024-10-01 20:25:54.971484] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:26:59.753 [2024-10-01 20:25:54.971506] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:26:59.753 [2024-10-01 20:25:54.974424] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:26:59.753 [2024-10-01 20:25:54.974478] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:26:59.753 BaseBdev2 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.753 20:25:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:00.011 spare_malloc 00:27:00.011 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.011 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:27:00.011 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.011 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:00.011 spare_delay 00:27:00.011 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.011 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:27:00.011 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.011 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:00.011 [2024-10-01 20:25:55.031920] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:27:00.011 [2024-10-01 20:25:55.032010] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:00.011 [2024-10-01 20:25:55.032044] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:27:00.011 [2024-10-01 20:25:55.032065] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:00.011 [2024-10-01 20:25:55.035011] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:00.011 [2024-10-01 20:25:55.035065] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:27:00.011 spare 00:27:00.011 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.011 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:27:00.011 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.011 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:00.011 [2024-10-01 20:25:55.040051] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:00.011 [2024-10-01 20:25:55.042585] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:00.011 [2024-10-01 20:25:55.042854] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:27:00.012 [2024-10-01 20:25:55.042880] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:27:00.012 [2024-10-01 20:25:55.043218] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:27:00.012 [2024-10-01 20:25:55.043436] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:27:00.012 [2024-10-01 20:25:55.043453] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:27:00.012 [2024-10-01 20:25:55.043639] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:00.012 "name": "raid_bdev1", 00:27:00.012 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:00.012 "strip_size_kb": 0, 00:27:00.012 "state": "online", 00:27:00.012 "raid_level": "raid1", 00:27:00.012 "superblock": true, 00:27:00.012 "num_base_bdevs": 2, 00:27:00.012 "num_base_bdevs_discovered": 2, 00:27:00.012 "num_base_bdevs_operational": 2, 00:27:00.012 "base_bdevs_list": [ 00:27:00.012 { 00:27:00.012 "name": "BaseBdev1", 00:27:00.012 "uuid": "a2181e12-4459-5331-81a1-85e2685fac47", 00:27:00.012 "is_configured": true, 00:27:00.012 "data_offset": 2048, 00:27:00.012 "data_size": 63488 00:27:00.012 }, 00:27:00.012 { 00:27:00.012 "name": "BaseBdev2", 00:27:00.012 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:00.012 "is_configured": true, 00:27:00.012 "data_offset": 2048, 00:27:00.012 "data_size": 63488 00:27:00.012 } 00:27:00.012 ] 00:27:00.012 }' 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:00.012 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:00.579 [2024-10-01 20:25:55.596573] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:00.579 [2024-10-01 20:25:55.696208] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:00.579 "name": "raid_bdev1", 00:27:00.579 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:00.579 "strip_size_kb": 0, 00:27:00.579 "state": "online", 00:27:00.579 "raid_level": "raid1", 00:27:00.579 "superblock": true, 00:27:00.579 "num_base_bdevs": 2, 00:27:00.579 "num_base_bdevs_discovered": 1, 00:27:00.579 "num_base_bdevs_operational": 1, 00:27:00.579 "base_bdevs_list": [ 00:27:00.579 { 00:27:00.579 "name": null, 00:27:00.579 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:00.579 "is_configured": false, 00:27:00.579 "data_offset": 0, 00:27:00.579 "data_size": 63488 00:27:00.579 }, 00:27:00.579 { 00:27:00.579 "name": "BaseBdev2", 00:27:00.579 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:00.579 "is_configured": true, 00:27:00.579 "data_offset": 2048, 00:27:00.579 "data_size": 63488 00:27:00.579 } 00:27:00.579 ] 00:27:00.579 }' 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:00.579 20:25:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:00.838 [2024-10-01 20:25:55.832386] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:27:00.838 I/O size of 3145728 is greater than zero copy threshold (65536). 00:27:00.838 Zero copy mechanism will not be used. 00:27:00.838 Running I/O for 60 seconds... 00:27:01.096 20:25:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:27:01.096 20:25:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:01.096 20:25:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:01.096 [2024-10-01 20:25:56.242210] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:27:01.096 20:25:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:01.096 20:25:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:27:01.096 [2024-10-01 20:25:56.310034] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:27:01.096 [2024-10-01 20:25:56.312676] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:27:01.355 [2024-10-01 20:25:56.423423] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:27:01.355 [2024-10-01 20:25:56.424360] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:27:01.614 [2024-10-01 20:25:56.645825] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:27:01.614 [2024-10-01 20:25:56.646382] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:27:01.871 191.00 IOPS, 573.00 MiB/s [2024-10-01 20:25:56.979828] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:27:02.130 [2024-10-01 20:25:57.206557] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:27:02.130 [2024-10-01 20:25:57.207238] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:27:02.130 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:02.130 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:02.130 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:02.130 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:02.130 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:02.130 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:02.130 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:02.130 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:02.130 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:02.130 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:02.130 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:02.130 "name": "raid_bdev1", 00:27:02.130 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:02.130 "strip_size_kb": 0, 00:27:02.130 "state": "online", 00:27:02.130 "raid_level": "raid1", 00:27:02.130 "superblock": true, 00:27:02.130 "num_base_bdevs": 2, 00:27:02.130 "num_base_bdevs_discovered": 2, 00:27:02.130 "num_base_bdevs_operational": 2, 00:27:02.130 "process": { 00:27:02.130 "type": "rebuild", 00:27:02.130 "target": "spare", 00:27:02.130 "progress": { 00:27:02.130 "blocks": 10240, 00:27:02.130 "percent": 16 00:27:02.130 } 00:27:02.130 }, 00:27:02.130 "base_bdevs_list": [ 00:27:02.130 { 00:27:02.130 "name": "spare", 00:27:02.130 "uuid": "42b41a81-a696-517e-b696-113fcfd70d0c", 00:27:02.130 "is_configured": true, 00:27:02.130 "data_offset": 2048, 00:27:02.130 "data_size": 63488 00:27:02.130 }, 00:27:02.130 { 00:27:02.130 "name": "BaseBdev2", 00:27:02.130 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:02.130 "is_configured": true, 00:27:02.130 "data_offset": 2048, 00:27:02.130 "data_size": 63488 00:27:02.130 } 00:27:02.130 ] 00:27:02.130 }' 00:27:02.130 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:02.389 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:02.389 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:02.389 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:27:02.389 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:27:02.389 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:02.389 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:02.389 [2024-10-01 20:25:57.449612] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:02.389 [2024-10-01 20:25:57.548734] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:27:02.647 [2024-10-01 20:25:57.658943] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:27:02.647 [2024-10-01 20:25:57.669550] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:02.647 [2024-10-01 20:25:57.669598] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:02.647 [2024-10-01 20:25:57.669621] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:27:02.647 [2024-10-01 20:25:57.715478] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006080 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:02.647 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:02.647 "name": "raid_bdev1", 00:27:02.647 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:02.647 "strip_size_kb": 0, 00:27:02.647 "state": "online", 00:27:02.647 "raid_level": "raid1", 00:27:02.647 "superblock": true, 00:27:02.647 "num_base_bdevs": 2, 00:27:02.647 "num_base_bdevs_discovered": 1, 00:27:02.647 "num_base_bdevs_operational": 1, 00:27:02.647 "base_bdevs_list": [ 00:27:02.647 { 00:27:02.647 "name": null, 00:27:02.647 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:02.647 "is_configured": false, 00:27:02.647 "data_offset": 0, 00:27:02.647 "data_size": 63488 00:27:02.647 }, 00:27:02.647 { 00:27:02.647 "name": "BaseBdev2", 00:27:02.647 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:02.647 "is_configured": true, 00:27:02.647 "data_offset": 2048, 00:27:02.647 "data_size": 63488 00:27:02.648 } 00:27:02.648 ] 00:27:02.648 }' 00:27:02.648 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:02.648 20:25:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:03.224 125.00 IOPS, 375.00 MiB/s 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:27:03.224 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:03.224 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:27:03.224 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:27:03.224 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:03.224 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:03.224 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:03.224 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:03.224 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:03.224 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:03.224 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:03.224 "name": "raid_bdev1", 00:27:03.224 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:03.224 "strip_size_kb": 0, 00:27:03.224 "state": "online", 00:27:03.224 "raid_level": "raid1", 00:27:03.224 "superblock": true, 00:27:03.224 "num_base_bdevs": 2, 00:27:03.224 "num_base_bdevs_discovered": 1, 00:27:03.224 "num_base_bdevs_operational": 1, 00:27:03.224 "base_bdevs_list": [ 00:27:03.224 { 00:27:03.224 "name": null, 00:27:03.224 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:03.224 "is_configured": false, 00:27:03.224 "data_offset": 0, 00:27:03.224 "data_size": 63488 00:27:03.224 }, 00:27:03.224 { 00:27:03.224 "name": "BaseBdev2", 00:27:03.224 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:03.224 "is_configured": true, 00:27:03.224 "data_offset": 2048, 00:27:03.224 "data_size": 63488 00:27:03.224 } 00:27:03.224 ] 00:27:03.224 }' 00:27:03.225 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:03.225 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:27:03.225 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:03.225 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:27:03.225 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:27:03.225 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:03.225 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:03.225 [2024-10-01 20:25:58.404923] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:27:03.225 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:03.225 20:25:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:27:03.522 [2024-10-01 20:25:58.472304] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:27:03.522 [2024-10-01 20:25:58.474994] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:27:03.522 [2024-10-01 20:25:58.607480] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:27:03.522 [2024-10-01 20:25:58.608156] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:27:03.780 [2024-10-01 20:25:58.818790] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:27:03.781 [2024-10-01 20:25:58.819190] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:27:04.038 142.00 IOPS, 426.00 MiB/s [2024-10-01 20:25:59.168789] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:27:04.038 [2024-10-01 20:25:59.169478] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:27:04.296 [2024-10-01 20:25:59.401108] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:27:04.296 [2024-10-01 20:25:59.401507] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:27:04.296 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:04.296 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:04.296 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:04.296 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:04.296 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:04.296 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:04.296 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:04.296 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:04.296 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:04.296 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:04.296 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:04.296 "name": "raid_bdev1", 00:27:04.296 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:04.296 "strip_size_kb": 0, 00:27:04.296 "state": "online", 00:27:04.296 "raid_level": "raid1", 00:27:04.296 "superblock": true, 00:27:04.296 "num_base_bdevs": 2, 00:27:04.296 "num_base_bdevs_discovered": 2, 00:27:04.296 "num_base_bdevs_operational": 2, 00:27:04.296 "process": { 00:27:04.296 "type": "rebuild", 00:27:04.296 "target": "spare", 00:27:04.296 "progress": { 00:27:04.296 "blocks": 10240, 00:27:04.296 "percent": 16 00:27:04.296 } 00:27:04.296 }, 00:27:04.296 "base_bdevs_list": [ 00:27:04.296 { 00:27:04.296 "name": "spare", 00:27:04.296 "uuid": "42b41a81-a696-517e-b696-113fcfd70d0c", 00:27:04.296 "is_configured": true, 00:27:04.296 "data_offset": 2048, 00:27:04.296 "data_size": 63488 00:27:04.296 }, 00:27:04.296 { 00:27:04.296 "name": "BaseBdev2", 00:27:04.296 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:04.296 "is_configured": true, 00:27:04.296 "data_offset": 2048, 00:27:04.296 "data_size": 63488 00:27:04.296 } 00:27:04.296 ] 00:27:04.296 }' 00:27:04.296 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:04.555 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:04.555 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:04.555 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:27:04.555 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:27:04.555 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:27:04.555 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:27:04.555 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:27:04.555 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:27:04.555 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:27:04.555 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=498 00:27:04.555 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:27:04.556 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:04.556 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:04.556 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:04.556 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:04.556 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:04.556 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:04.556 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:04.556 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:04.556 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:04.556 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:04.556 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:04.556 "name": "raid_bdev1", 00:27:04.556 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:04.556 "strip_size_kb": 0, 00:27:04.556 "state": "online", 00:27:04.556 "raid_level": "raid1", 00:27:04.556 "superblock": true, 00:27:04.556 "num_base_bdevs": 2, 00:27:04.556 "num_base_bdevs_discovered": 2, 00:27:04.556 "num_base_bdevs_operational": 2, 00:27:04.556 "process": { 00:27:04.556 "type": "rebuild", 00:27:04.556 "target": "spare", 00:27:04.556 "progress": { 00:27:04.556 "blocks": 10240, 00:27:04.556 "percent": 16 00:27:04.556 } 00:27:04.556 }, 00:27:04.556 "base_bdevs_list": [ 00:27:04.556 { 00:27:04.556 "name": "spare", 00:27:04.556 "uuid": "42b41a81-a696-517e-b696-113fcfd70d0c", 00:27:04.556 "is_configured": true, 00:27:04.556 "data_offset": 2048, 00:27:04.556 "data_size": 63488 00:27:04.556 }, 00:27:04.556 { 00:27:04.556 "name": "BaseBdev2", 00:27:04.556 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:04.556 "is_configured": true, 00:27:04.556 "data_offset": 2048, 00:27:04.556 "data_size": 63488 00:27:04.556 } 00:27:04.556 ] 00:27:04.556 }' 00:27:04.556 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:04.556 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:04.556 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:04.556 [2024-10-01 20:25:59.759691] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:27:04.556 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:27:04.556 20:25:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:27:04.815 118.50 IOPS, 355.50 MiB/s [2024-10-01 20:25:59.904407] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:27:05.380 [2024-10-01 20:26:00.400581] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:27:05.380 [2024-10-01 20:26:00.604568] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:27:05.638 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:27:05.638 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:05.638 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:05.638 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:05.638 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:05.638 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:05.638 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:05.638 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:05.638 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:05.638 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:05.638 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:05.638 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:05.638 "name": "raid_bdev1", 00:27:05.638 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:05.638 "strip_size_kb": 0, 00:27:05.638 "state": "online", 00:27:05.638 "raid_level": "raid1", 00:27:05.638 "superblock": true, 00:27:05.638 "num_base_bdevs": 2, 00:27:05.638 "num_base_bdevs_discovered": 2, 00:27:05.638 "num_base_bdevs_operational": 2, 00:27:05.638 "process": { 00:27:05.638 "type": "rebuild", 00:27:05.638 "target": "spare", 00:27:05.638 "progress": { 00:27:05.638 "blocks": 28672, 00:27:05.638 "percent": 45 00:27:05.638 } 00:27:05.638 }, 00:27:05.638 "base_bdevs_list": [ 00:27:05.638 { 00:27:05.638 "name": "spare", 00:27:05.638 "uuid": "42b41a81-a696-517e-b696-113fcfd70d0c", 00:27:05.638 "is_configured": true, 00:27:05.638 "data_offset": 2048, 00:27:05.638 "data_size": 63488 00:27:05.638 }, 00:27:05.638 { 00:27:05.638 "name": "BaseBdev2", 00:27:05.638 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:05.638 "is_configured": true, 00:27:05.638 "data_offset": 2048, 00:27:05.638 "data_size": 63488 00:27:05.638 } 00:27:05.638 ] 00:27:05.638 }' 00:27:05.638 107.40 IOPS, 322.20 MiB/s 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:05.898 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:05.898 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:05.898 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:27:05.898 20:26:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:27:05.898 [2024-10-01 20:26:01.070864] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:27:06.465 [2024-10-01 20:26:01.422599] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:27:06.465 [2024-10-01 20:26:01.423313] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:27:06.465 [2024-10-01 20:26:01.650812] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:27:06.465 [2024-10-01 20:26:01.651439] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:27:06.723 94.33 IOPS, 283.00 MiB/s 20:26:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:27:06.724 20:26:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:06.724 20:26:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:06.724 20:26:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:06.724 20:26:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:06.724 20:26:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:06.724 20:26:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:06.724 20:26:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:06.724 20:26:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:06.724 20:26:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:06.982 [2024-10-01 20:26:01.988654] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 4 20:26:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:06.982 3008 offset_end: 49152 00:27:06.982 [2024-10-01 20:26:01.989557] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:27:06.982 20:26:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:06.982 "name": "raid_bdev1", 00:27:06.982 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:06.982 "strip_size_kb": 0, 00:27:06.982 "state": "online", 00:27:06.982 "raid_level": "raid1", 00:27:06.982 "superblock": true, 00:27:06.982 "num_base_bdevs": 2, 00:27:06.982 "num_base_bdevs_discovered": 2, 00:27:06.982 "num_base_bdevs_operational": 2, 00:27:06.982 "process": { 00:27:06.982 "type": "rebuild", 00:27:06.982 "target": "spare", 00:27:06.982 "progress": { 00:27:06.982 "blocks": 43008, 00:27:06.982 "percent": 67 00:27:06.982 } 00:27:06.982 }, 00:27:06.982 "base_bdevs_list": [ 00:27:06.982 { 00:27:06.982 "name": "spare", 00:27:06.982 "uuid": "42b41a81-a696-517e-b696-113fcfd70d0c", 00:27:06.982 "is_configured": true, 00:27:06.982 "data_offset": 2048, 00:27:06.982 "data_size": 63488 00:27:06.982 }, 00:27:06.982 { 00:27:06.982 "name": "BaseBdev2", 00:27:06.982 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:06.982 "is_configured": true, 00:27:06.982 "data_offset": 2048, 00:27:06.982 "data_size": 63488 00:27:06.982 } 00:27:06.982 ] 00:27:06.982 }' 00:27:06.982 20:26:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:06.983 20:26:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:06.983 20:26:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:06.983 20:26:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:27:06.983 20:26:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:27:06.983 [2024-10-01 20:26:02.226913] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:27:07.919 86.00 IOPS, 258.00 MiB/s [2024-10-01 20:26:02.931128] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:27:07.919 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:27:07.919 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:07.919 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:07.919 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:07.919 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:07.919 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:07.919 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:07.919 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:07.919 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:07.919 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:07.919 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:07.919 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:07.919 "name": "raid_bdev1", 00:27:07.919 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:07.919 "strip_size_kb": 0, 00:27:07.919 "state": "online", 00:27:07.919 "raid_level": "raid1", 00:27:07.919 "superblock": true, 00:27:07.919 "num_base_bdevs": 2, 00:27:07.919 "num_base_bdevs_discovered": 2, 00:27:07.919 "num_base_bdevs_operational": 2, 00:27:07.919 "process": { 00:27:07.919 "type": "rebuild", 00:27:07.919 "target": "spare", 00:27:07.919 "progress": { 00:27:07.919 "blocks": 59392, 00:27:07.919 "percent": 93 00:27:07.919 } 00:27:07.919 }, 00:27:07.919 "base_bdevs_list": [ 00:27:07.919 { 00:27:07.919 "name": "spare", 00:27:07.919 "uuid": "42b41a81-a696-517e-b696-113fcfd70d0c", 00:27:07.919 "is_configured": true, 00:27:07.919 "data_offset": 2048, 00:27:07.919 "data_size": 63488 00:27:07.919 }, 00:27:07.919 { 00:27:07.919 "name": "BaseBdev2", 00:27:07.919 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:07.919 "is_configured": true, 00:27:07.919 "data_offset": 2048, 00:27:07.919 "data_size": 63488 00:27:07.919 } 00:27:07.919 ] 00:27:07.919 }' 00:27:07.919 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:08.178 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:08.178 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:08.178 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:27:08.178 20:26:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:27:08.178 [2024-10-01 20:26:03.275207] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:27:08.178 [2024-10-01 20:26:03.375202] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:27:08.178 [2024-10-01 20:26:03.378327] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:09.029 79.25 IOPS, 237.75 MiB/s 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:27:09.029 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:09.029 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:09.029 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:09.029 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:09.029 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:09.289 "name": "raid_bdev1", 00:27:09.289 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:09.289 "strip_size_kb": 0, 00:27:09.289 "state": "online", 00:27:09.289 "raid_level": "raid1", 00:27:09.289 "superblock": true, 00:27:09.289 "num_base_bdevs": 2, 00:27:09.289 "num_base_bdevs_discovered": 2, 00:27:09.289 "num_base_bdevs_operational": 2, 00:27:09.289 "base_bdevs_list": [ 00:27:09.289 { 00:27:09.289 "name": "spare", 00:27:09.289 "uuid": "42b41a81-a696-517e-b696-113fcfd70d0c", 00:27:09.289 "is_configured": true, 00:27:09.289 "data_offset": 2048, 00:27:09.289 "data_size": 63488 00:27:09.289 }, 00:27:09.289 { 00:27:09.289 "name": "BaseBdev2", 00:27:09.289 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:09.289 "is_configured": true, 00:27:09.289 "data_offset": 2048, 00:27:09.289 "data_size": 63488 00:27:09.289 } 00:27:09.289 ] 00:27:09.289 }' 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:09.289 "name": "raid_bdev1", 00:27:09.289 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:09.289 "strip_size_kb": 0, 00:27:09.289 "state": "online", 00:27:09.289 "raid_level": "raid1", 00:27:09.289 "superblock": true, 00:27:09.289 "num_base_bdevs": 2, 00:27:09.289 "num_base_bdevs_discovered": 2, 00:27:09.289 "num_base_bdevs_operational": 2, 00:27:09.289 "base_bdevs_list": [ 00:27:09.289 { 00:27:09.289 "name": "spare", 00:27:09.289 "uuid": "42b41a81-a696-517e-b696-113fcfd70d0c", 00:27:09.289 "is_configured": true, 00:27:09.289 "data_offset": 2048, 00:27:09.289 "data_size": 63488 00:27:09.289 }, 00:27:09.289 { 00:27:09.289 "name": "BaseBdev2", 00:27:09.289 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:09.289 "is_configured": true, 00:27:09.289 "data_offset": 2048, 00:27:09.289 "data_size": 63488 00:27:09.289 } 00:27:09.289 ] 00:27:09.289 }' 00:27:09.289 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:09.549 "name": "raid_bdev1", 00:27:09.549 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:09.549 "strip_size_kb": 0, 00:27:09.549 "state": "online", 00:27:09.549 "raid_level": "raid1", 00:27:09.549 "superblock": true, 00:27:09.549 "num_base_bdevs": 2, 00:27:09.549 "num_base_bdevs_discovered": 2, 00:27:09.549 "num_base_bdevs_operational": 2, 00:27:09.549 "base_bdevs_list": [ 00:27:09.549 { 00:27:09.549 "name": "spare", 00:27:09.549 "uuid": "42b41a81-a696-517e-b696-113fcfd70d0c", 00:27:09.549 "is_configured": true, 00:27:09.549 "data_offset": 2048, 00:27:09.549 "data_size": 63488 00:27:09.549 }, 00:27:09.549 { 00:27:09.549 "name": "BaseBdev2", 00:27:09.549 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:09.549 "is_configured": true, 00:27:09.549 "data_offset": 2048, 00:27:09.549 "data_size": 63488 00:27:09.549 } 00:27:09.549 ] 00:27:09.549 }' 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:09.549 20:26:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:10.066 73.44 IOPS, 220.33 MiB/s 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:27:10.066 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.066 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:10.066 [2024-10-01 20:26:05.171098] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:10.066 [2024-10-01 20:26:05.171157] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:10.066 00:27:10.066 Latency(us) 00:27:10.066 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:10.066 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:27:10.066 raid_bdev1 : 9.42 71.31 213.93 0.00 0.00 18682.35 296.03 125829.12 00:27:10.066 =================================================================================================================== 00:27:10.066 Total : 71.31 213.93 0.00 0.00 18682.35 296.03 125829.12 00:27:10.066 { 00:27:10.066 "results": [ 00:27:10.066 { 00:27:10.066 "job": "raid_bdev1", 00:27:10.066 "core_mask": "0x1", 00:27:10.067 "workload": "randrw", 00:27:10.067 "percentage": 50, 00:27:10.067 "status": "finished", 00:27:10.067 "queue_depth": 2, 00:27:10.067 "io_size": 3145728, 00:27:10.067 "runtime": 9.423858, 00:27:10.067 "iops": 71.30837497763655, 00:27:10.067 "mibps": 213.92512493290963, 00:27:10.067 "io_failed": 0, 00:27:10.067 "io_timeout": 0, 00:27:10.067 "avg_latency_us": 18682.35082251082, 00:27:10.067 "min_latency_us": 296.0290909090909, 00:27:10.067 "max_latency_us": 125829.12 00:27:10.067 } 00:27:10.067 ], 00:27:10.067 "core_count": 1 00:27:10.067 } 00:27:10.067 [2024-10-01 20:26:05.279870] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:10.067 [2024-10-01 20:26:05.279957] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:10.067 [2024-10-01 20:26:05.280078] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:10.067 [2024-10-01 20:26:05.280103] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:27:10.067 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.067 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:10.067 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:27:10.067 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.067 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:10.067 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.326 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:27:10.326 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:27:10.326 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:27:10.326 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:27:10.326 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:27:10.326 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:27:10.326 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:27:10.326 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:27:10.326 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:27:10.326 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:27:10.326 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:27:10.326 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:27:10.326 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:27:10.585 /dev/nbd0 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:27:10.585 1+0 records in 00:27:10.585 1+0 records out 00:27:10.585 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000411551 s, 10.0 MB/s 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:27:10.585 20:26:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:27:10.844 /dev/nbd1 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:27:10.844 1+0 records in 00:27:10.844 1+0 records out 00:27:10.844 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000375088 s, 10.9 MB/s 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:27:10.844 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:27:11.103 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:27:11.103 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:27:11.103 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:27:11.103 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:27:11.103 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:27:11.103 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:27:11.103 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:27:11.362 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:27:11.362 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:27:11.362 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:27:11.362 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:27:11.362 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:27:11.362 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:27:11.362 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:27:11.362 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:27:11.362 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:27:11.362 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:27:11.362 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:27:11.362 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:27:11.362 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:27:11.362 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:27:11.362 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:11.621 [2024-10-01 20:26:06.863907] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:27:11.621 [2024-10-01 20:26:06.864010] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:11.621 [2024-10-01 20:26:06.864050] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:27:11.621 [2024-10-01 20:26:06.864072] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:11.621 [2024-10-01 20:26:06.867397] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:11.621 [2024-10-01 20:26:06.867462] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:27:11.621 [2024-10-01 20:26:06.867583] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:27:11.621 [2024-10-01 20:26:06.867665] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:27:11.621 [2024-10-01 20:26:06.867917] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:11.621 spare 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.621 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:11.880 [2024-10-01 20:26:06.968118] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:27:11.880 [2024-10-01 20:26:06.968437] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:27:11.880 [2024-10-01 20:26:06.968982] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b0d0 00:27:11.880 [2024-10-01 20:26:06.969254] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:27:11.880 [2024-10-01 20:26:06.969308] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:27:11.880 [2024-10-01 20:26:06.969562] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:11.880 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.880 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:27:11.880 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:11.880 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:11.880 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:11.880 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:11.880 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:27:11.880 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:11.880 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:11.880 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:11.880 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:11.880 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:11.880 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.880 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:11.880 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:11.880 20:26:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.880 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:11.880 "name": "raid_bdev1", 00:27:11.880 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:11.880 "strip_size_kb": 0, 00:27:11.880 "state": "online", 00:27:11.880 "raid_level": "raid1", 00:27:11.880 "superblock": true, 00:27:11.880 "num_base_bdevs": 2, 00:27:11.880 "num_base_bdevs_discovered": 2, 00:27:11.880 "num_base_bdevs_operational": 2, 00:27:11.880 "base_bdevs_list": [ 00:27:11.880 { 00:27:11.880 "name": "spare", 00:27:11.880 "uuid": "42b41a81-a696-517e-b696-113fcfd70d0c", 00:27:11.880 "is_configured": true, 00:27:11.880 "data_offset": 2048, 00:27:11.880 "data_size": 63488 00:27:11.880 }, 00:27:11.880 { 00:27:11.880 "name": "BaseBdev2", 00:27:11.880 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:11.880 "is_configured": true, 00:27:11.880 "data_offset": 2048, 00:27:11.880 "data_size": 63488 00:27:11.880 } 00:27:11.880 ] 00:27:11.880 }' 00:27:11.880 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:11.880 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:12.447 "name": "raid_bdev1", 00:27:12.447 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:12.447 "strip_size_kb": 0, 00:27:12.447 "state": "online", 00:27:12.447 "raid_level": "raid1", 00:27:12.447 "superblock": true, 00:27:12.447 "num_base_bdevs": 2, 00:27:12.447 "num_base_bdevs_discovered": 2, 00:27:12.447 "num_base_bdevs_operational": 2, 00:27:12.447 "base_bdevs_list": [ 00:27:12.447 { 00:27:12.447 "name": "spare", 00:27:12.447 "uuid": "42b41a81-a696-517e-b696-113fcfd70d0c", 00:27:12.447 "is_configured": true, 00:27:12.447 "data_offset": 2048, 00:27:12.447 "data_size": 63488 00:27:12.447 }, 00:27:12.447 { 00:27:12.447 "name": "BaseBdev2", 00:27:12.447 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:12.447 "is_configured": true, 00:27:12.447 "data_offset": 2048, 00:27:12.447 "data_size": 63488 00:27:12.447 } 00:27:12.447 ] 00:27:12.447 }' 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:12.447 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:12.706 [2024-10-01 20:26:07.712481] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:12.706 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:12.706 "name": "raid_bdev1", 00:27:12.706 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:12.706 "strip_size_kb": 0, 00:27:12.706 "state": "online", 00:27:12.706 "raid_level": "raid1", 00:27:12.706 "superblock": true, 00:27:12.706 "num_base_bdevs": 2, 00:27:12.706 "num_base_bdevs_discovered": 1, 00:27:12.706 "num_base_bdevs_operational": 1, 00:27:12.706 "base_bdevs_list": [ 00:27:12.706 { 00:27:12.706 "name": null, 00:27:12.706 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:12.706 "is_configured": false, 00:27:12.706 "data_offset": 0, 00:27:12.706 "data_size": 63488 00:27:12.706 }, 00:27:12.706 { 00:27:12.706 "name": "BaseBdev2", 00:27:12.706 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:12.706 "is_configured": true, 00:27:12.706 "data_offset": 2048, 00:27:12.706 "data_size": 63488 00:27:12.706 } 00:27:12.706 ] 00:27:12.707 }' 00:27:12.707 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:12.707 20:26:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:13.280 20:26:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:27:13.280 20:26:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:13.280 20:26:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:13.280 [2024-10-01 20:26:08.232757] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:27:13.280 [2024-10-01 20:26:08.233033] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:27:13.280 [2024-10-01 20:26:08.233057] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:27:13.280 [2024-10-01 20:26:08.233181] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:27:13.280 [2024-10-01 20:26:08.249245] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b1a0 00:27:13.280 20:26:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:13.280 20:26:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:27:13.280 [2024-10-01 20:26:08.251947] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:27:14.215 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:14.215 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:14.215 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:14.215 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:14.215 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:14.215 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:14.215 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:14.215 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:14.215 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:14.215 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:14.215 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:14.215 "name": "raid_bdev1", 00:27:14.216 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:14.216 "strip_size_kb": 0, 00:27:14.216 "state": "online", 00:27:14.216 "raid_level": "raid1", 00:27:14.216 "superblock": true, 00:27:14.216 "num_base_bdevs": 2, 00:27:14.216 "num_base_bdevs_discovered": 2, 00:27:14.216 "num_base_bdevs_operational": 2, 00:27:14.216 "process": { 00:27:14.216 "type": "rebuild", 00:27:14.216 "target": "spare", 00:27:14.216 "progress": { 00:27:14.216 "blocks": 20480, 00:27:14.216 "percent": 32 00:27:14.216 } 00:27:14.216 }, 00:27:14.216 "base_bdevs_list": [ 00:27:14.216 { 00:27:14.216 "name": "spare", 00:27:14.216 "uuid": "42b41a81-a696-517e-b696-113fcfd70d0c", 00:27:14.216 "is_configured": true, 00:27:14.216 "data_offset": 2048, 00:27:14.216 "data_size": 63488 00:27:14.216 }, 00:27:14.216 { 00:27:14.216 "name": "BaseBdev2", 00:27:14.216 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:14.216 "is_configured": true, 00:27:14.216 "data_offset": 2048, 00:27:14.216 "data_size": 63488 00:27:14.216 } 00:27:14.216 ] 00:27:14.216 }' 00:27:14.216 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:14.216 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:14.216 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:14.216 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:27:14.216 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:27:14.216 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:14.216 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:14.216 [2024-10-01 20:26:09.429560] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:14.216 [2024-10-01 20:26:09.461472] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:27:14.216 [2024-10-01 20:26:09.461558] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:14.216 [2024-10-01 20:26:09.461588] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:14.216 [2024-10-01 20:26:09.461601] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:14.480 "name": "raid_bdev1", 00:27:14.480 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:14.480 "strip_size_kb": 0, 00:27:14.480 "state": "online", 00:27:14.480 "raid_level": "raid1", 00:27:14.480 "superblock": true, 00:27:14.480 "num_base_bdevs": 2, 00:27:14.480 "num_base_bdevs_discovered": 1, 00:27:14.480 "num_base_bdevs_operational": 1, 00:27:14.480 "base_bdevs_list": [ 00:27:14.480 { 00:27:14.480 "name": null, 00:27:14.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:14.480 "is_configured": false, 00:27:14.480 "data_offset": 0, 00:27:14.480 "data_size": 63488 00:27:14.480 }, 00:27:14.480 { 00:27:14.480 "name": "BaseBdev2", 00:27:14.480 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:14.480 "is_configured": true, 00:27:14.480 "data_offset": 2048, 00:27:14.480 "data_size": 63488 00:27:14.480 } 00:27:14.480 ] 00:27:14.480 }' 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:14.480 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:14.739 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:27:14.739 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:14.739 20:26:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:14.739 [2024-10-01 20:26:09.983703] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:27:14.739 [2024-10-01 20:26:09.983958] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:14.739 [2024-10-01 20:26:09.984141] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:27:14.739 [2024-10-01 20:26:09.984265] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:14.739 [2024-10-01 20:26:09.985054] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:14.739 [2024-10-01 20:26:09.985090] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:27:14.739 [2024-10-01 20:26:09.985238] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:27:14.739 [2024-10-01 20:26:09.985261] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:27:14.739 [2024-10-01 20:26:09.985307] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:27:14.739 [2024-10-01 20:26:09.985339] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:27:14.998 [2024-10-01 20:26:10.000291] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b270 00:27:14.998 spare 00:27:14.998 20:26:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:14.998 20:26:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:27:14.998 [2024-10-01 20:26:10.002873] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:15.935 "name": "raid_bdev1", 00:27:15.935 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:15.935 "strip_size_kb": 0, 00:27:15.935 "state": "online", 00:27:15.935 "raid_level": "raid1", 00:27:15.935 "superblock": true, 00:27:15.935 "num_base_bdevs": 2, 00:27:15.935 "num_base_bdevs_discovered": 2, 00:27:15.935 "num_base_bdevs_operational": 2, 00:27:15.935 "process": { 00:27:15.935 "type": "rebuild", 00:27:15.935 "target": "spare", 00:27:15.935 "progress": { 00:27:15.935 "blocks": 20480, 00:27:15.935 "percent": 32 00:27:15.935 } 00:27:15.935 }, 00:27:15.935 "base_bdevs_list": [ 00:27:15.935 { 00:27:15.935 "name": "spare", 00:27:15.935 "uuid": "42b41a81-a696-517e-b696-113fcfd70d0c", 00:27:15.935 "is_configured": true, 00:27:15.935 "data_offset": 2048, 00:27:15.935 "data_size": 63488 00:27:15.935 }, 00:27:15.935 { 00:27:15.935 "name": "BaseBdev2", 00:27:15.935 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:15.935 "is_configured": true, 00:27:15.935 "data_offset": 2048, 00:27:15.935 "data_size": 63488 00:27:15.935 } 00:27:15.935 ] 00:27:15.935 }' 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:15.935 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:15.935 [2024-10-01 20:26:11.176485] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:16.194 [2024-10-01 20:26:11.212309] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:27:16.194 [2024-10-01 20:26:11.212446] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:16.194 [2024-10-01 20:26:11.212473] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:16.194 [2024-10-01 20:26:11.212491] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:16.194 "name": "raid_bdev1", 00:27:16.194 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:16.194 "strip_size_kb": 0, 00:27:16.194 "state": "online", 00:27:16.194 "raid_level": "raid1", 00:27:16.194 "superblock": true, 00:27:16.194 "num_base_bdevs": 2, 00:27:16.194 "num_base_bdevs_discovered": 1, 00:27:16.194 "num_base_bdevs_operational": 1, 00:27:16.194 "base_bdevs_list": [ 00:27:16.194 { 00:27:16.194 "name": null, 00:27:16.194 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:16.194 "is_configured": false, 00:27:16.194 "data_offset": 0, 00:27:16.194 "data_size": 63488 00:27:16.194 }, 00:27:16.194 { 00:27:16.194 "name": "BaseBdev2", 00:27:16.194 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:16.194 "is_configured": true, 00:27:16.194 "data_offset": 2048, 00:27:16.194 "data_size": 63488 00:27:16.194 } 00:27:16.194 ] 00:27:16.194 }' 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:16.194 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:16.761 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:27:16.761 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:16.761 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:27:16.761 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:27:16.761 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:16.761 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:16.761 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:16.761 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:16.761 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:16.761 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:16.761 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:16.761 "name": "raid_bdev1", 00:27:16.761 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:16.761 "strip_size_kb": 0, 00:27:16.761 "state": "online", 00:27:16.761 "raid_level": "raid1", 00:27:16.761 "superblock": true, 00:27:16.762 "num_base_bdevs": 2, 00:27:16.762 "num_base_bdevs_discovered": 1, 00:27:16.762 "num_base_bdevs_operational": 1, 00:27:16.762 "base_bdevs_list": [ 00:27:16.762 { 00:27:16.762 "name": null, 00:27:16.762 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:16.762 "is_configured": false, 00:27:16.762 "data_offset": 0, 00:27:16.762 "data_size": 63488 00:27:16.762 }, 00:27:16.762 { 00:27:16.762 "name": "BaseBdev2", 00:27:16.762 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:16.762 "is_configured": true, 00:27:16.762 "data_offset": 2048, 00:27:16.762 "data_size": 63488 00:27:16.762 } 00:27:16.762 ] 00:27:16.762 }' 00:27:16.762 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:16.762 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:27:16.762 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:16.762 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:27:16.762 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:27:16.762 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:16.762 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:16.762 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:16.762 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:27:16.762 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:16.762 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:16.762 [2024-10-01 20:26:11.969897] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:27:16.762 [2024-10-01 20:26:11.969980] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:16.762 [2024-10-01 20:26:11.970024] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:27:16.762 [2024-10-01 20:26:11.970048] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:16.762 [2024-10-01 20:26:11.970623] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:16.762 [2024-10-01 20:26:11.970665] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:27:16.762 [2024-10-01 20:26:11.970789] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:27:16.762 [2024-10-01 20:26:11.970822] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:27:16.762 [2024-10-01 20:26:11.970840] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:27:16.762 [2024-10-01 20:26:11.970861] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:27:16.762 BaseBdev1 00:27:16.762 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:16.762 20:26:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:27:18.135 20:26:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:18.135 20:26:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:18.135 20:26:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:18.135 20:26:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:18.135 20:26:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:18.135 20:26:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:27:18.135 20:26:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:18.135 20:26:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:18.135 20:26:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:18.135 20:26:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:18.135 20:26:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:18.135 20:26:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:18.135 20:26:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:18.135 20:26:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:18.135 20:26:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:18.135 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:18.135 "name": "raid_bdev1", 00:27:18.135 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:18.135 "strip_size_kb": 0, 00:27:18.135 "state": "online", 00:27:18.135 "raid_level": "raid1", 00:27:18.135 "superblock": true, 00:27:18.135 "num_base_bdevs": 2, 00:27:18.135 "num_base_bdevs_discovered": 1, 00:27:18.135 "num_base_bdevs_operational": 1, 00:27:18.135 "base_bdevs_list": [ 00:27:18.135 { 00:27:18.135 "name": null, 00:27:18.135 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:18.135 "is_configured": false, 00:27:18.135 "data_offset": 0, 00:27:18.135 "data_size": 63488 00:27:18.135 }, 00:27:18.135 { 00:27:18.135 "name": "BaseBdev2", 00:27:18.135 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:18.135 "is_configured": true, 00:27:18.135 "data_offset": 2048, 00:27:18.136 "data_size": 63488 00:27:18.136 } 00:27:18.136 ] 00:27:18.136 }' 00:27:18.136 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:18.136 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:18.393 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:27:18.393 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:18.393 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:27:18.393 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:27:18.393 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:18.393 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:18.393 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:18.393 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:18.393 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:18.393 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:18.393 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:18.393 "name": "raid_bdev1", 00:27:18.393 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:18.393 "strip_size_kb": 0, 00:27:18.393 "state": "online", 00:27:18.393 "raid_level": "raid1", 00:27:18.393 "superblock": true, 00:27:18.393 "num_base_bdevs": 2, 00:27:18.393 "num_base_bdevs_discovered": 1, 00:27:18.393 "num_base_bdevs_operational": 1, 00:27:18.393 "base_bdevs_list": [ 00:27:18.393 { 00:27:18.393 "name": null, 00:27:18.393 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:18.393 "is_configured": false, 00:27:18.393 "data_offset": 0, 00:27:18.393 "data_size": 63488 00:27:18.393 }, 00:27:18.393 { 00:27:18.393 "name": "BaseBdev2", 00:27:18.393 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:18.394 "is_configured": true, 00:27:18.394 "data_offset": 2048, 00:27:18.394 "data_size": 63488 00:27:18.394 } 00:27:18.394 ] 00:27:18.394 }' 00:27:18.394 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:18.394 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:27:18.394 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@650 -- # local es=0 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:18.652 [2024-10-01 20:26:13.682674] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:18.652 [2024-10-01 20:26:13.683013] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:27:18.652 [2024-10-01 20:26:13.683042] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:27:18.652 request: 00:27:18.652 { 00:27:18.652 "base_bdev": "BaseBdev1", 00:27:18.652 "raid_bdev": "raid_bdev1", 00:27:18.652 "method": "bdev_raid_add_base_bdev", 00:27:18.652 "req_id": 1 00:27:18.652 } 00:27:18.652 Got JSON-RPC error response 00:27:18.652 response: 00:27:18.652 { 00:27:18.652 "code": -22, 00:27:18.652 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:27:18.652 } 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # es=1 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:27:18.652 20:26:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:19.603 "name": "raid_bdev1", 00:27:19.603 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:19.603 "strip_size_kb": 0, 00:27:19.603 "state": "online", 00:27:19.603 "raid_level": "raid1", 00:27:19.603 "superblock": true, 00:27:19.603 "num_base_bdevs": 2, 00:27:19.603 "num_base_bdevs_discovered": 1, 00:27:19.603 "num_base_bdevs_operational": 1, 00:27:19.603 "base_bdevs_list": [ 00:27:19.603 { 00:27:19.603 "name": null, 00:27:19.603 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:19.603 "is_configured": false, 00:27:19.603 "data_offset": 0, 00:27:19.603 "data_size": 63488 00:27:19.603 }, 00:27:19.603 { 00:27:19.603 "name": "BaseBdev2", 00:27:19.603 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:19.603 "is_configured": true, 00:27:19.603 "data_offset": 2048, 00:27:19.603 "data_size": 63488 00:27:19.603 } 00:27:19.603 ] 00:27:19.603 }' 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:19.603 20:26:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:20.168 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:27:20.168 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:20.168 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:27:20.168 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:27:20.168 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:20.168 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:20.168 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:20.168 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:20.168 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:20.168 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:20.168 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:20.168 "name": "raid_bdev1", 00:27:20.168 "uuid": "327f80b5-0119-442f-881d-271daaa6364c", 00:27:20.168 "strip_size_kb": 0, 00:27:20.168 "state": "online", 00:27:20.168 "raid_level": "raid1", 00:27:20.168 "superblock": true, 00:27:20.168 "num_base_bdevs": 2, 00:27:20.168 "num_base_bdevs_discovered": 1, 00:27:20.168 "num_base_bdevs_operational": 1, 00:27:20.168 "base_bdevs_list": [ 00:27:20.168 { 00:27:20.168 "name": null, 00:27:20.168 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:20.168 "is_configured": false, 00:27:20.168 "data_offset": 0, 00:27:20.168 "data_size": 63488 00:27:20.168 }, 00:27:20.168 { 00:27:20.168 "name": "BaseBdev2", 00:27:20.168 "uuid": "a8df5829-7a9f-51ad-9ee5-ca0fc465b1c9", 00:27:20.168 "is_configured": true, 00:27:20.168 "data_offset": 2048, 00:27:20.168 "data_size": 63488 00:27:20.168 } 00:27:20.168 ] 00:27:20.168 }' 00:27:20.168 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:20.168 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:27:20.168 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:20.168 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:27:20.168 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 77867 00:27:20.169 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@950 -- # '[' -z 77867 ']' 00:27:20.169 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # kill -0 77867 00:27:20.169 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@955 -- # uname 00:27:20.169 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:27:20.169 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 77867 00:27:20.169 killing process with pid 77867 00:27:20.169 Received shutdown signal, test time was about 19.554682 seconds 00:27:20.169 00:27:20.169 Latency(us) 00:27:20.169 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:20.169 =================================================================================================================== 00:27:20.169 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:27:20.169 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:27:20.169 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:27:20.169 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@968 -- # echo 'killing process with pid 77867' 00:27:20.169 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@969 -- # kill 77867 00:27:20.169 [2024-10-01 20:26:15.389892] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:20.169 20:26:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@974 -- # wait 77867 00:27:20.169 [2024-10-01 20:26:15.390066] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:20.169 [2024-10-01 20:26:15.390152] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:20.169 [2024-10-01 20:26:15.390171] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:27:20.428 [2024-10-01 20:26:15.594536] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:22.329 ************************************ 00:27:22.329 END TEST raid_rebuild_test_sb_io 00:27:22.329 ************************************ 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:27:22.329 00:27:22.329 real 0m23.845s 00:27:22.329 user 0m31.853s 00:27:22.329 sys 0m2.224s 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:27:22.329 20:26:17 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:27:22.329 20:26:17 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 4 false false true 00:27:22.329 20:26:17 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:27:22.329 20:26:17 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:27:22.329 20:26:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:22.329 ************************************ 00:27:22.329 START TEST raid_rebuild_test 00:27:22.329 ************************************ 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 4 false false true 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=78599 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 78599 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@831 -- # '[' -z 78599 ']' 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:22.329 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:27:22.329 20:26:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:22.587 I/O size of 3145728 is greater than zero copy threshold (65536). 00:27:22.588 Zero copy mechanism will not be used. 00:27:22.588 [2024-10-01 20:26:17.647215] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:27:22.588 [2024-10-01 20:26:17.647419] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78599 ] 00:27:22.588 [2024-10-01 20:26:17.825659] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:22.851 [2024-10-01 20:26:18.092361] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:27:23.109 [2024-10-01 20:26:18.298035] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:23.109 [2024-10-01 20:26:18.298081] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # return 0 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:23.674 BaseBdev1_malloc 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:23.674 [2024-10-01 20:26:18.756040] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:27:23.674 [2024-10-01 20:26:18.756277] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:23.674 [2024-10-01 20:26:18.756493] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:27:23.674 [2024-10-01 20:26:18.756678] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:23.674 [2024-10-01 20:26:18.759995] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:23.674 [2024-10-01 20:26:18.760054] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:27:23.674 BaseBdev1 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:23.674 BaseBdev2_malloc 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.674 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:23.674 [2024-10-01 20:26:18.812526] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:27:23.674 [2024-10-01 20:26:18.812783] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:23.674 [2024-10-01 20:26:18.813001] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:27:23.675 [2024-10-01 20:26:18.813051] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:23.675 [2024-10-01 20:26:18.816184] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:23.675 [2024-10-01 20:26:18.816244] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:27:23.675 BaseBdev2 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:23.675 BaseBdev3_malloc 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:23.675 [2024-10-01 20:26:18.865007] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:27:23.675 [2024-10-01 20:26:18.865225] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:23.675 [2024-10-01 20:26:18.865439] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:27:23.675 [2024-10-01 20:26:18.865604] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:23.675 [2024-10-01 20:26:18.868748] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:23.675 [2024-10-01 20:26:18.868806] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:27:23.675 BaseBdev3 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:23.675 BaseBdev4_malloc 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:23.675 [2024-10-01 20:26:18.921580] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:27:23.675 [2024-10-01 20:26:18.921666] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:23.675 [2024-10-01 20:26:18.921734] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:27:23.675 [2024-10-01 20:26:18.921770] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:23.675 [2024-10-01 20:26:18.924854] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:23.675 [2024-10-01 20:26:18.924916] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:27:23.675 BaseBdev4 00:27:23.675 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.934 20:26:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:27:23.934 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.934 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:23.934 spare_malloc 00:27:23.934 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.934 20:26:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:27:23.934 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.934 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:23.934 spare_delay 00:27:23.934 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.934 20:26:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:27:23.934 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.934 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:23.934 [2024-10-01 20:26:18.990087] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:27:23.934 [2024-10-01 20:26:18.990323] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:23.934 [2024-10-01 20:26:18.990497] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:27:23.935 [2024-10-01 20:26:18.990551] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:23.935 [2024-10-01 20:26:18.993756] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:23.935 [2024-10-01 20:26:18.993813] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:27:23.935 spare 00:27:23.935 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.935 20:26:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:27:23.935 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.935 20:26:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:23.935 [2024-10-01 20:26:19.002193] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:23.935 [2024-10-01 20:26:19.004869] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:23.935 [2024-10-01 20:26:19.005137] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:27:23.935 [2024-10-01 20:26:19.005414] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:27:23.935 [2024-10-01 20:26:19.005785] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:27:23.935 [2024-10-01 20:26:19.005936] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:27:23.935 [2024-10-01 20:26:19.006508] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:27:23.935 [2024-10-01 20:26:19.006957] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:27:23.935 [2024-10-01 20:26:19.006985] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:27:23.935 [2024-10-01 20:26:19.007241] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:23.935 "name": "raid_bdev1", 00:27:23.935 "uuid": "c83cff15-88d6-49f2-936a-54baa313caf6", 00:27:23.935 "strip_size_kb": 0, 00:27:23.935 "state": "online", 00:27:23.935 "raid_level": "raid1", 00:27:23.935 "superblock": false, 00:27:23.935 "num_base_bdevs": 4, 00:27:23.935 "num_base_bdevs_discovered": 4, 00:27:23.935 "num_base_bdevs_operational": 4, 00:27:23.935 "base_bdevs_list": [ 00:27:23.935 { 00:27:23.935 "name": "BaseBdev1", 00:27:23.935 "uuid": "f30a17ca-a1ab-5b6f-adf9-9a9373ae934d", 00:27:23.935 "is_configured": true, 00:27:23.935 "data_offset": 0, 00:27:23.935 "data_size": 65536 00:27:23.935 }, 00:27:23.935 { 00:27:23.935 "name": "BaseBdev2", 00:27:23.935 "uuid": "e3df3bc6-ea08-5d66-bace-628dd4aed5d1", 00:27:23.935 "is_configured": true, 00:27:23.935 "data_offset": 0, 00:27:23.935 "data_size": 65536 00:27:23.935 }, 00:27:23.935 { 00:27:23.935 "name": "BaseBdev3", 00:27:23.935 "uuid": "c6e61cbf-2f3f-5918-8bb4-92b65ad00274", 00:27:23.935 "is_configured": true, 00:27:23.935 "data_offset": 0, 00:27:23.935 "data_size": 65536 00:27:23.935 }, 00:27:23.935 { 00:27:23.935 "name": "BaseBdev4", 00:27:23.935 "uuid": "96d8add6-95e2-548e-b470-6249c53ddba2", 00:27:23.935 "is_configured": true, 00:27:23.935 "data_offset": 0, 00:27:23.935 "data_size": 65536 00:27:23.935 } 00:27:23.935 ] 00:27:23.935 }' 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:23.935 20:26:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:27:24.502 [2024-10-01 20:26:19.535796] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:27:24.502 20:26:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:27:24.760 [2024-10-01 20:26:19.955555] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:27:24.760 /dev/nbd0 00:27:24.760 20:26:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:27:24.760 20:26:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:27:24.760 20:26:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:27:24.760 20:26:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:27:24.760 20:26:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:27:24.760 20:26:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:27:24.760 20:26:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:27:24.760 20:26:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:27:24.760 20:26:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:27:24.760 20:26:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:27:24.760 20:26:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:27:24.760 1+0 records in 00:27:24.760 1+0 records out 00:27:24.760 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000365454 s, 11.2 MB/s 00:27:24.760 20:26:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:27:24.760 20:26:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:27:24.760 20:26:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:27:25.018 20:26:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:27:25.019 20:26:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:27:25.019 20:26:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:27:25.019 20:26:20 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:27:25.019 20:26:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:27:25.019 20:26:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:27:25.019 20:26:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:27:34.986 65536+0 records in 00:27:34.986 65536+0 records out 00:27:34.986 33554432 bytes (34 MB, 32 MiB) copied, 8.78558 s, 3.8 MB/s 00:27:34.986 20:26:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:27:34.986 20:26:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:27:34.986 20:26:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:27:34.986 20:26:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:27:34.986 20:26:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:27:34.986 20:26:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:27:34.986 20:26:28 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:27:34.986 [2024-10-01 20:26:29.109310] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:34.986 20:26:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:27:34.986 20:26:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:27:34.986 20:26:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:27:34.986 20:26:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:27:34.986 20:26:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:27:34.986 20:26:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:27:34.986 20:26:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:27:34.986 20:26:29 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:27:34.986 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:27:34.986 20:26:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:34.986 20:26:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.986 [2024-10-01 20:26:29.145404] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:27:34.986 20:26:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:34.986 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:34.987 "name": "raid_bdev1", 00:27:34.987 "uuid": "c83cff15-88d6-49f2-936a-54baa313caf6", 00:27:34.987 "strip_size_kb": 0, 00:27:34.987 "state": "online", 00:27:34.987 "raid_level": "raid1", 00:27:34.987 "superblock": false, 00:27:34.987 "num_base_bdevs": 4, 00:27:34.987 "num_base_bdevs_discovered": 3, 00:27:34.987 "num_base_bdevs_operational": 3, 00:27:34.987 "base_bdevs_list": [ 00:27:34.987 { 00:27:34.987 "name": null, 00:27:34.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:34.987 "is_configured": false, 00:27:34.987 "data_offset": 0, 00:27:34.987 "data_size": 65536 00:27:34.987 }, 00:27:34.987 { 00:27:34.987 "name": "BaseBdev2", 00:27:34.987 "uuid": "e3df3bc6-ea08-5d66-bace-628dd4aed5d1", 00:27:34.987 "is_configured": true, 00:27:34.987 "data_offset": 0, 00:27:34.987 "data_size": 65536 00:27:34.987 }, 00:27:34.987 { 00:27:34.987 "name": "BaseBdev3", 00:27:34.987 "uuid": "c6e61cbf-2f3f-5918-8bb4-92b65ad00274", 00:27:34.987 "is_configured": true, 00:27:34.987 "data_offset": 0, 00:27:34.987 "data_size": 65536 00:27:34.987 }, 00:27:34.987 { 00:27:34.987 "name": "BaseBdev4", 00:27:34.987 "uuid": "96d8add6-95e2-548e-b470-6249c53ddba2", 00:27:34.987 "is_configured": true, 00:27:34.987 "data_offset": 0, 00:27:34.987 "data_size": 65536 00:27:34.987 } 00:27:34.987 ] 00:27:34.987 }' 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:34.987 [2024-10-01 20:26:29.657581] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:27:34.987 [2024-10-01 20:26:29.671347] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09d70 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:34.987 20:26:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:27:34.987 [2024-10-01 20:26:29.673966] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:27:35.552 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:35.552 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:35.552 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:35.552 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:35.552 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:35.552 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:35.552 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:35.552 20:26:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:35.552 20:26:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:35.552 20:26:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:35.552 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:35.552 "name": "raid_bdev1", 00:27:35.552 "uuid": "c83cff15-88d6-49f2-936a-54baa313caf6", 00:27:35.552 "strip_size_kb": 0, 00:27:35.552 "state": "online", 00:27:35.552 "raid_level": "raid1", 00:27:35.552 "superblock": false, 00:27:35.552 "num_base_bdevs": 4, 00:27:35.552 "num_base_bdevs_discovered": 4, 00:27:35.552 "num_base_bdevs_operational": 4, 00:27:35.552 "process": { 00:27:35.552 "type": "rebuild", 00:27:35.552 "target": "spare", 00:27:35.552 "progress": { 00:27:35.552 "blocks": 20480, 00:27:35.552 "percent": 31 00:27:35.552 } 00:27:35.552 }, 00:27:35.552 "base_bdevs_list": [ 00:27:35.552 { 00:27:35.552 "name": "spare", 00:27:35.552 "uuid": "db6dbc50-9c92-5e33-a904-a128ad85ef60", 00:27:35.552 "is_configured": true, 00:27:35.552 "data_offset": 0, 00:27:35.552 "data_size": 65536 00:27:35.552 }, 00:27:35.552 { 00:27:35.552 "name": "BaseBdev2", 00:27:35.552 "uuid": "e3df3bc6-ea08-5d66-bace-628dd4aed5d1", 00:27:35.553 "is_configured": true, 00:27:35.553 "data_offset": 0, 00:27:35.553 "data_size": 65536 00:27:35.553 }, 00:27:35.553 { 00:27:35.553 "name": "BaseBdev3", 00:27:35.553 "uuid": "c6e61cbf-2f3f-5918-8bb4-92b65ad00274", 00:27:35.553 "is_configured": true, 00:27:35.553 "data_offset": 0, 00:27:35.553 "data_size": 65536 00:27:35.553 }, 00:27:35.553 { 00:27:35.553 "name": "BaseBdev4", 00:27:35.553 "uuid": "96d8add6-95e2-548e-b470-6249c53ddba2", 00:27:35.553 "is_configured": true, 00:27:35.553 "data_offset": 0, 00:27:35.553 "data_size": 65536 00:27:35.553 } 00:27:35.553 ] 00:27:35.553 }' 00:27:35.553 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:35.553 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:35.553 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:35.811 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:27:35.811 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:27:35.811 20:26:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:35.811 20:26:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:35.811 [2024-10-01 20:26:30.843708] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:35.812 [2024-10-01 20:26:30.883131] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:27:35.812 [2024-10-01 20:26:30.883296] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:35.812 [2024-10-01 20:26:30.883325] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:35.812 [2024-10-01 20:26:30.883343] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:35.812 "name": "raid_bdev1", 00:27:35.812 "uuid": "c83cff15-88d6-49f2-936a-54baa313caf6", 00:27:35.812 "strip_size_kb": 0, 00:27:35.812 "state": "online", 00:27:35.812 "raid_level": "raid1", 00:27:35.812 "superblock": false, 00:27:35.812 "num_base_bdevs": 4, 00:27:35.812 "num_base_bdevs_discovered": 3, 00:27:35.812 "num_base_bdevs_operational": 3, 00:27:35.812 "base_bdevs_list": [ 00:27:35.812 { 00:27:35.812 "name": null, 00:27:35.812 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:35.812 "is_configured": false, 00:27:35.812 "data_offset": 0, 00:27:35.812 "data_size": 65536 00:27:35.812 }, 00:27:35.812 { 00:27:35.812 "name": "BaseBdev2", 00:27:35.812 "uuid": "e3df3bc6-ea08-5d66-bace-628dd4aed5d1", 00:27:35.812 "is_configured": true, 00:27:35.812 "data_offset": 0, 00:27:35.812 "data_size": 65536 00:27:35.812 }, 00:27:35.812 { 00:27:35.812 "name": "BaseBdev3", 00:27:35.812 "uuid": "c6e61cbf-2f3f-5918-8bb4-92b65ad00274", 00:27:35.812 "is_configured": true, 00:27:35.812 "data_offset": 0, 00:27:35.812 "data_size": 65536 00:27:35.812 }, 00:27:35.812 { 00:27:35.812 "name": "BaseBdev4", 00:27:35.812 "uuid": "96d8add6-95e2-548e-b470-6249c53ddba2", 00:27:35.812 "is_configured": true, 00:27:35.812 "data_offset": 0, 00:27:35.812 "data_size": 65536 00:27:35.812 } 00:27:35.812 ] 00:27:35.812 }' 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:35.812 20:26:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:36.377 20:26:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:27:36.377 20:26:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:36.377 20:26:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:27:36.377 20:26:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:27:36.377 20:26:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:36.377 20:26:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:36.377 20:26:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:36.377 20:26:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:36.377 20:26:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:36.377 20:26:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:36.377 20:26:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:36.377 "name": "raid_bdev1", 00:27:36.377 "uuid": "c83cff15-88d6-49f2-936a-54baa313caf6", 00:27:36.377 "strip_size_kb": 0, 00:27:36.377 "state": "online", 00:27:36.377 "raid_level": "raid1", 00:27:36.378 "superblock": false, 00:27:36.378 "num_base_bdevs": 4, 00:27:36.378 "num_base_bdevs_discovered": 3, 00:27:36.378 "num_base_bdevs_operational": 3, 00:27:36.378 "base_bdevs_list": [ 00:27:36.378 { 00:27:36.378 "name": null, 00:27:36.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:36.378 "is_configured": false, 00:27:36.378 "data_offset": 0, 00:27:36.378 "data_size": 65536 00:27:36.378 }, 00:27:36.378 { 00:27:36.378 "name": "BaseBdev2", 00:27:36.378 "uuid": "e3df3bc6-ea08-5d66-bace-628dd4aed5d1", 00:27:36.378 "is_configured": true, 00:27:36.378 "data_offset": 0, 00:27:36.378 "data_size": 65536 00:27:36.378 }, 00:27:36.378 { 00:27:36.378 "name": "BaseBdev3", 00:27:36.378 "uuid": "c6e61cbf-2f3f-5918-8bb4-92b65ad00274", 00:27:36.378 "is_configured": true, 00:27:36.378 "data_offset": 0, 00:27:36.378 "data_size": 65536 00:27:36.378 }, 00:27:36.378 { 00:27:36.378 "name": "BaseBdev4", 00:27:36.378 "uuid": "96d8add6-95e2-548e-b470-6249c53ddba2", 00:27:36.378 "is_configured": true, 00:27:36.378 "data_offset": 0, 00:27:36.378 "data_size": 65536 00:27:36.378 } 00:27:36.378 ] 00:27:36.378 }' 00:27:36.378 20:26:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:36.378 20:26:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:27:36.378 20:26:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:36.378 20:26:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:27:36.378 20:26:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:27:36.378 20:26:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:36.378 20:26:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:36.378 [2024-10-01 20:26:31.581775] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:27:36.378 [2024-10-01 20:26:31.594905] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d09e40 00:27:36.378 20:26:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:36.378 20:26:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:27:36.378 [2024-10-01 20:26:31.597487] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:27:37.753 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:37.753 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:37.753 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:37.753 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:37.753 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:37.753 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:37.753 20:26:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:37.753 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:37.753 20:26:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:37.753 20:26:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:37.753 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:37.753 "name": "raid_bdev1", 00:27:37.753 "uuid": "c83cff15-88d6-49f2-936a-54baa313caf6", 00:27:37.753 "strip_size_kb": 0, 00:27:37.753 "state": "online", 00:27:37.753 "raid_level": "raid1", 00:27:37.753 "superblock": false, 00:27:37.753 "num_base_bdevs": 4, 00:27:37.753 "num_base_bdevs_discovered": 4, 00:27:37.753 "num_base_bdevs_operational": 4, 00:27:37.753 "process": { 00:27:37.753 "type": "rebuild", 00:27:37.753 "target": "spare", 00:27:37.753 "progress": { 00:27:37.753 "blocks": 20480, 00:27:37.753 "percent": 31 00:27:37.753 } 00:27:37.753 }, 00:27:37.754 "base_bdevs_list": [ 00:27:37.754 { 00:27:37.754 "name": "spare", 00:27:37.754 "uuid": "db6dbc50-9c92-5e33-a904-a128ad85ef60", 00:27:37.754 "is_configured": true, 00:27:37.754 "data_offset": 0, 00:27:37.754 "data_size": 65536 00:27:37.754 }, 00:27:37.754 { 00:27:37.754 "name": "BaseBdev2", 00:27:37.754 "uuid": "e3df3bc6-ea08-5d66-bace-628dd4aed5d1", 00:27:37.754 "is_configured": true, 00:27:37.754 "data_offset": 0, 00:27:37.754 "data_size": 65536 00:27:37.754 }, 00:27:37.754 { 00:27:37.754 "name": "BaseBdev3", 00:27:37.754 "uuid": "c6e61cbf-2f3f-5918-8bb4-92b65ad00274", 00:27:37.754 "is_configured": true, 00:27:37.754 "data_offset": 0, 00:27:37.754 "data_size": 65536 00:27:37.754 }, 00:27:37.754 { 00:27:37.754 "name": "BaseBdev4", 00:27:37.754 "uuid": "96d8add6-95e2-548e-b470-6249c53ddba2", 00:27:37.754 "is_configured": true, 00:27:37.754 "data_offset": 0, 00:27:37.754 "data_size": 65536 00:27:37.754 } 00:27:37.754 ] 00:27:37.754 }' 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:37.754 [2024-10-01 20:26:32.771090] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:27:37.754 [2024-10-01 20:26:32.806450] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000d09e40 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:37.754 "name": "raid_bdev1", 00:27:37.754 "uuid": "c83cff15-88d6-49f2-936a-54baa313caf6", 00:27:37.754 "strip_size_kb": 0, 00:27:37.754 "state": "online", 00:27:37.754 "raid_level": "raid1", 00:27:37.754 "superblock": false, 00:27:37.754 "num_base_bdevs": 4, 00:27:37.754 "num_base_bdevs_discovered": 3, 00:27:37.754 "num_base_bdevs_operational": 3, 00:27:37.754 "process": { 00:27:37.754 "type": "rebuild", 00:27:37.754 "target": "spare", 00:27:37.754 "progress": { 00:27:37.754 "blocks": 24576, 00:27:37.754 "percent": 37 00:27:37.754 } 00:27:37.754 }, 00:27:37.754 "base_bdevs_list": [ 00:27:37.754 { 00:27:37.754 "name": "spare", 00:27:37.754 "uuid": "db6dbc50-9c92-5e33-a904-a128ad85ef60", 00:27:37.754 "is_configured": true, 00:27:37.754 "data_offset": 0, 00:27:37.754 "data_size": 65536 00:27:37.754 }, 00:27:37.754 { 00:27:37.754 "name": null, 00:27:37.754 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:37.754 "is_configured": false, 00:27:37.754 "data_offset": 0, 00:27:37.754 "data_size": 65536 00:27:37.754 }, 00:27:37.754 { 00:27:37.754 "name": "BaseBdev3", 00:27:37.754 "uuid": "c6e61cbf-2f3f-5918-8bb4-92b65ad00274", 00:27:37.754 "is_configured": true, 00:27:37.754 "data_offset": 0, 00:27:37.754 "data_size": 65536 00:27:37.754 }, 00:27:37.754 { 00:27:37.754 "name": "BaseBdev4", 00:27:37.754 "uuid": "96d8add6-95e2-548e-b470-6249c53ddba2", 00:27:37.754 "is_configured": true, 00:27:37.754 "data_offset": 0, 00:27:37.754 "data_size": 65536 00:27:37.754 } 00:27:37.754 ] 00:27:37.754 }' 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=531 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:37.754 20:26:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:38.013 20:26:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:38.013 "name": "raid_bdev1", 00:27:38.013 "uuid": "c83cff15-88d6-49f2-936a-54baa313caf6", 00:27:38.013 "strip_size_kb": 0, 00:27:38.013 "state": "online", 00:27:38.013 "raid_level": "raid1", 00:27:38.013 "superblock": false, 00:27:38.013 "num_base_bdevs": 4, 00:27:38.013 "num_base_bdevs_discovered": 3, 00:27:38.013 "num_base_bdevs_operational": 3, 00:27:38.013 "process": { 00:27:38.013 "type": "rebuild", 00:27:38.013 "target": "spare", 00:27:38.013 "progress": { 00:27:38.013 "blocks": 26624, 00:27:38.013 "percent": 40 00:27:38.013 } 00:27:38.013 }, 00:27:38.013 "base_bdevs_list": [ 00:27:38.013 { 00:27:38.013 "name": "spare", 00:27:38.013 "uuid": "db6dbc50-9c92-5e33-a904-a128ad85ef60", 00:27:38.013 "is_configured": true, 00:27:38.013 "data_offset": 0, 00:27:38.013 "data_size": 65536 00:27:38.013 }, 00:27:38.013 { 00:27:38.013 "name": null, 00:27:38.013 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:38.013 "is_configured": false, 00:27:38.013 "data_offset": 0, 00:27:38.013 "data_size": 65536 00:27:38.013 }, 00:27:38.013 { 00:27:38.013 "name": "BaseBdev3", 00:27:38.013 "uuid": "c6e61cbf-2f3f-5918-8bb4-92b65ad00274", 00:27:38.013 "is_configured": true, 00:27:38.013 "data_offset": 0, 00:27:38.013 "data_size": 65536 00:27:38.013 }, 00:27:38.013 { 00:27:38.013 "name": "BaseBdev4", 00:27:38.013 "uuid": "96d8add6-95e2-548e-b470-6249c53ddba2", 00:27:38.013 "is_configured": true, 00:27:38.013 "data_offset": 0, 00:27:38.013 "data_size": 65536 00:27:38.013 } 00:27:38.013 ] 00:27:38.013 }' 00:27:38.013 20:26:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:38.013 20:26:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:38.013 20:26:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:38.013 20:26:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:27:38.013 20:26:33 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:27:38.948 20:26:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:27:38.948 20:26:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:38.948 20:26:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:38.948 20:26:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:38.948 20:26:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:38.948 20:26:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:38.948 20:26:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:38.948 20:26:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:38.948 20:26:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:38.948 20:26:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:38.948 20:26:34 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:38.948 20:26:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:38.948 "name": "raid_bdev1", 00:27:38.948 "uuid": "c83cff15-88d6-49f2-936a-54baa313caf6", 00:27:38.948 "strip_size_kb": 0, 00:27:38.948 "state": "online", 00:27:38.948 "raid_level": "raid1", 00:27:38.948 "superblock": false, 00:27:38.948 "num_base_bdevs": 4, 00:27:38.948 "num_base_bdevs_discovered": 3, 00:27:38.948 "num_base_bdevs_operational": 3, 00:27:38.948 "process": { 00:27:38.948 "type": "rebuild", 00:27:38.949 "target": "spare", 00:27:38.949 "progress": { 00:27:38.949 "blocks": 51200, 00:27:38.949 "percent": 78 00:27:38.949 } 00:27:38.949 }, 00:27:38.949 "base_bdevs_list": [ 00:27:38.949 { 00:27:38.949 "name": "spare", 00:27:38.949 "uuid": "db6dbc50-9c92-5e33-a904-a128ad85ef60", 00:27:38.949 "is_configured": true, 00:27:38.949 "data_offset": 0, 00:27:38.949 "data_size": 65536 00:27:38.949 }, 00:27:38.949 { 00:27:38.949 "name": null, 00:27:38.949 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:38.949 "is_configured": false, 00:27:38.949 "data_offset": 0, 00:27:38.949 "data_size": 65536 00:27:38.949 }, 00:27:38.949 { 00:27:38.949 "name": "BaseBdev3", 00:27:38.949 "uuid": "c6e61cbf-2f3f-5918-8bb4-92b65ad00274", 00:27:38.949 "is_configured": true, 00:27:38.949 "data_offset": 0, 00:27:38.949 "data_size": 65536 00:27:38.949 }, 00:27:38.949 { 00:27:38.949 "name": "BaseBdev4", 00:27:38.949 "uuid": "96d8add6-95e2-548e-b470-6249c53ddba2", 00:27:38.949 "is_configured": true, 00:27:38.949 "data_offset": 0, 00:27:38.949 "data_size": 65536 00:27:38.949 } 00:27:38.949 ] 00:27:38.949 }' 00:27:38.949 20:26:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:39.208 20:26:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:39.208 20:26:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:39.208 20:26:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:27:39.208 20:26:34 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:27:39.852 [2024-10-01 20:26:34.822587] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:27:39.852 [2024-10-01 20:26:34.822929] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:27:39.852 [2024-10-01 20:26:34.823022] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:40.110 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:27:40.110 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:40.110 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:40.110 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:40.110 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:40.110 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:40.110 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:40.110 20:26:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:40.110 20:26:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:40.110 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:40.110 20:26:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:40.110 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:40.110 "name": "raid_bdev1", 00:27:40.110 "uuid": "c83cff15-88d6-49f2-936a-54baa313caf6", 00:27:40.110 "strip_size_kb": 0, 00:27:40.110 "state": "online", 00:27:40.110 "raid_level": "raid1", 00:27:40.110 "superblock": false, 00:27:40.110 "num_base_bdevs": 4, 00:27:40.110 "num_base_bdevs_discovered": 3, 00:27:40.110 "num_base_bdevs_operational": 3, 00:27:40.110 "base_bdevs_list": [ 00:27:40.110 { 00:27:40.110 "name": "spare", 00:27:40.110 "uuid": "db6dbc50-9c92-5e33-a904-a128ad85ef60", 00:27:40.110 "is_configured": true, 00:27:40.110 "data_offset": 0, 00:27:40.110 "data_size": 65536 00:27:40.110 }, 00:27:40.110 { 00:27:40.110 "name": null, 00:27:40.110 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:40.110 "is_configured": false, 00:27:40.110 "data_offset": 0, 00:27:40.110 "data_size": 65536 00:27:40.110 }, 00:27:40.110 { 00:27:40.110 "name": "BaseBdev3", 00:27:40.110 "uuid": "c6e61cbf-2f3f-5918-8bb4-92b65ad00274", 00:27:40.110 "is_configured": true, 00:27:40.110 "data_offset": 0, 00:27:40.110 "data_size": 65536 00:27:40.110 }, 00:27:40.110 { 00:27:40.110 "name": "BaseBdev4", 00:27:40.110 "uuid": "96d8add6-95e2-548e-b470-6249c53ddba2", 00:27:40.110 "is_configured": true, 00:27:40.110 "data_offset": 0, 00:27:40.110 "data_size": 65536 00:27:40.110 } 00:27:40.110 ] 00:27:40.110 }' 00:27:40.110 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:40.370 "name": "raid_bdev1", 00:27:40.370 "uuid": "c83cff15-88d6-49f2-936a-54baa313caf6", 00:27:40.370 "strip_size_kb": 0, 00:27:40.370 "state": "online", 00:27:40.370 "raid_level": "raid1", 00:27:40.370 "superblock": false, 00:27:40.370 "num_base_bdevs": 4, 00:27:40.370 "num_base_bdevs_discovered": 3, 00:27:40.370 "num_base_bdevs_operational": 3, 00:27:40.370 "base_bdevs_list": [ 00:27:40.370 { 00:27:40.370 "name": "spare", 00:27:40.370 "uuid": "db6dbc50-9c92-5e33-a904-a128ad85ef60", 00:27:40.370 "is_configured": true, 00:27:40.370 "data_offset": 0, 00:27:40.370 "data_size": 65536 00:27:40.370 }, 00:27:40.370 { 00:27:40.370 "name": null, 00:27:40.370 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:40.370 "is_configured": false, 00:27:40.370 "data_offset": 0, 00:27:40.370 "data_size": 65536 00:27:40.370 }, 00:27:40.370 { 00:27:40.370 "name": "BaseBdev3", 00:27:40.370 "uuid": "c6e61cbf-2f3f-5918-8bb4-92b65ad00274", 00:27:40.370 "is_configured": true, 00:27:40.370 "data_offset": 0, 00:27:40.370 "data_size": 65536 00:27:40.370 }, 00:27:40.370 { 00:27:40.370 "name": "BaseBdev4", 00:27:40.370 "uuid": "96d8add6-95e2-548e-b470-6249c53ddba2", 00:27:40.370 "is_configured": true, 00:27:40.370 "data_offset": 0, 00:27:40.370 "data_size": 65536 00:27:40.370 } 00:27:40.370 ] 00:27:40.370 }' 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:27:40.370 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:40.629 "name": "raid_bdev1", 00:27:40.629 "uuid": "c83cff15-88d6-49f2-936a-54baa313caf6", 00:27:40.629 "strip_size_kb": 0, 00:27:40.629 "state": "online", 00:27:40.629 "raid_level": "raid1", 00:27:40.629 "superblock": false, 00:27:40.629 "num_base_bdevs": 4, 00:27:40.629 "num_base_bdevs_discovered": 3, 00:27:40.629 "num_base_bdevs_operational": 3, 00:27:40.629 "base_bdevs_list": [ 00:27:40.629 { 00:27:40.629 "name": "spare", 00:27:40.629 "uuid": "db6dbc50-9c92-5e33-a904-a128ad85ef60", 00:27:40.629 "is_configured": true, 00:27:40.629 "data_offset": 0, 00:27:40.629 "data_size": 65536 00:27:40.629 }, 00:27:40.629 { 00:27:40.629 "name": null, 00:27:40.629 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:40.629 "is_configured": false, 00:27:40.629 "data_offset": 0, 00:27:40.629 "data_size": 65536 00:27:40.629 }, 00:27:40.629 { 00:27:40.629 "name": "BaseBdev3", 00:27:40.629 "uuid": "c6e61cbf-2f3f-5918-8bb4-92b65ad00274", 00:27:40.629 "is_configured": true, 00:27:40.629 "data_offset": 0, 00:27:40.629 "data_size": 65536 00:27:40.629 }, 00:27:40.629 { 00:27:40.629 "name": "BaseBdev4", 00:27:40.629 "uuid": "96d8add6-95e2-548e-b470-6249c53ddba2", 00:27:40.629 "is_configured": true, 00:27:40.629 "data_offset": 0, 00:27:40.629 "data_size": 65536 00:27:40.629 } 00:27:40.629 ] 00:27:40.629 }' 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:40.629 20:26:35 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:41.199 [2024-10-01 20:26:36.166063] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:27:41.199 [2024-10-01 20:26:36.166312] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:27:41.199 [2024-10-01 20:26:36.166543] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:27:41.199 [2024-10-01 20:26:36.166678] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:27:41.199 [2024-10-01 20:26:36.166697] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:27:41.199 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:27:41.458 /dev/nbd0 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:27:41.458 1+0 records in 00:27:41.458 1+0 records out 00:27:41.458 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000470436 s, 8.7 MB/s 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:27:41.458 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:27:41.717 /dev/nbd1 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:27:41.717 1+0 records in 00:27:41.717 1+0 records out 00:27:41.717 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000458106 s, 8.9 MB/s 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:27:41.717 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:27:41.718 20:26:36 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:27:41.718 20:26:36 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:27:41.976 20:26:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:27:41.976 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:27:41.976 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:27:41.976 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:27:41.976 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:27:41.976 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:27:41.976 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:27:42.235 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:27:42.235 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:27:42.235 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:27:42.235 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:27:42.235 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:27:42.235 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:27:42.235 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:27:42.235 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:27:42.235 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:27:42.235 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 78599 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@950 -- # '[' -z 78599 ']' 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # kill -0 78599 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@955 -- # uname 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 78599 00:27:42.495 killing process with pid 78599 00:27:42.495 Received shutdown signal, test time was about 60.000000 seconds 00:27:42.495 00:27:42.495 Latency(us) 00:27:42.495 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:42.495 =================================================================================================================== 00:27:42.495 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 78599' 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@969 -- # kill 78599 00:27:42.495 [2024-10-01 20:26:37.743498] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:27:42.495 20:26:37 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@974 -- # wait 78599 00:27:43.064 [2024-10-01 20:26:38.176305] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:27:44.968 20:26:40 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:27:44.968 00:27:44.968 real 0m22.501s 00:27:44.968 user 0m24.872s 00:27:44.968 sys 0m3.822s 00:27:44.968 20:26:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:27:44.968 ************************************ 00:27:44.968 END TEST raid_rebuild_test 00:27:44.968 ************************************ 00:27:44.968 20:26:40 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:27:44.968 20:26:40 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 4 true false true 00:27:44.968 20:26:40 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:27:44.969 20:26:40 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:27:44.969 20:26:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:27:44.969 ************************************ 00:27:44.969 START TEST raid_rebuild_test_sb 00:27:44.969 ************************************ 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 4 true false true 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=79094 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 79094 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@831 -- # '[' -z 79094 ']' 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:44.969 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:27:44.969 20:26:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:44.969 I/O size of 3145728 is greater than zero copy threshold (65536). 00:27:44.969 Zero copy mechanism will not be used. 00:27:44.969 [2024-10-01 20:26:40.211806] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:27:44.969 [2024-10-01 20:26:40.212016] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79094 ] 00:27:45.230 [2024-10-01 20:26:40.407030] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:45.492 [2024-10-01 20:26:40.654219] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:27:45.751 [2024-10-01 20:26:40.865647] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:45.751 [2024-10-01 20:26:40.865735] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:27:46.318 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:27:46.318 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # return 0 00:27:46.318 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:46.319 BaseBdev1_malloc 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:46.319 [2024-10-01 20:26:41.339278] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:27:46.319 [2024-10-01 20:26:41.339416] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:46.319 [2024-10-01 20:26:41.339455] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:27:46.319 [2024-10-01 20:26:41.339481] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:46.319 [2024-10-01 20:26:41.342611] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:46.319 [2024-10-01 20:26:41.342666] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:27:46.319 BaseBdev1 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:46.319 BaseBdev2_malloc 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:46.319 [2024-10-01 20:26:41.397927] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:27:46.319 [2024-10-01 20:26:41.398026] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:46.319 [2024-10-01 20:26:41.398081] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:27:46.319 [2024-10-01 20:26:41.398101] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:46.319 [2024-10-01 20:26:41.401131] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:46.319 [2024-10-01 20:26:41.401213] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:27:46.319 BaseBdev2 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:46.319 BaseBdev3_malloc 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:46.319 [2024-10-01 20:26:41.456411] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:27:46.319 [2024-10-01 20:26:41.456519] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:46.319 [2024-10-01 20:26:41.456555] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:27:46.319 [2024-10-01 20:26:41.456576] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:46.319 [2024-10-01 20:26:41.459835] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:46.319 [2024-10-01 20:26:41.460030] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:27:46.319 BaseBdev3 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:46.319 BaseBdev4_malloc 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:46.319 [2024-10-01 20:26:41.509860] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:27:46.319 [2024-10-01 20:26:41.509953] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:46.319 [2024-10-01 20:26:41.509986] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:27:46.319 [2024-10-01 20:26:41.510006] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:46.319 [2024-10-01 20:26:41.513066] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:46.319 [2024-10-01 20:26:41.513154] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:27:46.319 BaseBdev4 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:46.319 spare_malloc 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:46.319 spare_delay 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:46.319 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:46.577 [2024-10-01 20:26:41.575291] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:27:46.577 [2024-10-01 20:26:41.575393] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:27:46.577 [2024-10-01 20:26:41.575431] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:27:46.577 [2024-10-01 20:26:41.575452] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:27:46.577 [2024-10-01 20:26:41.578466] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:27:46.577 [2024-10-01 20:26:41.578653] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:27:46.577 spare 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:46.577 [2024-10-01 20:26:41.587412] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:27:46.577 [2024-10-01 20:26:41.590065] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:27:46.577 [2024-10-01 20:26:41.590185] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:27:46.577 [2024-10-01 20:26:41.590268] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:27:46.577 [2024-10-01 20:26:41.590564] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:27:46.577 [2024-10-01 20:26:41.590588] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:27:46.577 [2024-10-01 20:26:41.591018] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:27:46.577 [2024-10-01 20:26:41.591248] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:27:46.577 [2024-10-01 20:26:41.591385] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:27:46.577 [2024-10-01 20:26:41.591658] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:46.577 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:46.577 "name": "raid_bdev1", 00:27:46.577 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:27:46.577 "strip_size_kb": 0, 00:27:46.577 "state": "online", 00:27:46.577 "raid_level": "raid1", 00:27:46.577 "superblock": true, 00:27:46.577 "num_base_bdevs": 4, 00:27:46.577 "num_base_bdevs_discovered": 4, 00:27:46.577 "num_base_bdevs_operational": 4, 00:27:46.577 "base_bdevs_list": [ 00:27:46.577 { 00:27:46.577 "name": "BaseBdev1", 00:27:46.577 "uuid": "98049d2d-2248-58b4-a7b0-b478c41a6517", 00:27:46.577 "is_configured": true, 00:27:46.577 "data_offset": 2048, 00:27:46.578 "data_size": 63488 00:27:46.578 }, 00:27:46.578 { 00:27:46.578 "name": "BaseBdev2", 00:27:46.578 "uuid": "b51137bf-830c-5855-b862-e71a62b5d33f", 00:27:46.578 "is_configured": true, 00:27:46.578 "data_offset": 2048, 00:27:46.578 "data_size": 63488 00:27:46.578 }, 00:27:46.578 { 00:27:46.578 "name": "BaseBdev3", 00:27:46.578 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:27:46.578 "is_configured": true, 00:27:46.578 "data_offset": 2048, 00:27:46.578 "data_size": 63488 00:27:46.578 }, 00:27:46.578 { 00:27:46.578 "name": "BaseBdev4", 00:27:46.578 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:27:46.578 "is_configured": true, 00:27:46.578 "data_offset": 2048, 00:27:46.578 "data_size": 63488 00:27:46.578 } 00:27:46.578 ] 00:27:46.578 }' 00:27:46.578 20:26:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:46.578 20:26:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:46.836 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:27:46.836 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:27:46.836 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:46.836 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:46.836 [2024-10-01 20:26:42.072234] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:27:47.094 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:27:47.352 [2024-10-01 20:26:42.419956] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:27:47.352 /dev/nbd0 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:27:47.352 1+0 records in 00:27:47.352 1+0 records out 00:27:47.352 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000438856 s, 9.3 MB/s 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:27:47.352 20:26:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:27:57.335 63488+0 records in 00:27:57.335 63488+0 records out 00:27:57.335 32505856 bytes (33 MB, 31 MiB) copied, 8.61393 s, 3.8 MB/s 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:27:57.335 [2024-10-01 20:26:51.398792] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:57.335 [2024-10-01 20:26:51.430932] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:57.335 "name": "raid_bdev1", 00:27:57.335 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:27:57.335 "strip_size_kb": 0, 00:27:57.335 "state": "online", 00:27:57.335 "raid_level": "raid1", 00:27:57.335 "superblock": true, 00:27:57.335 "num_base_bdevs": 4, 00:27:57.335 "num_base_bdevs_discovered": 3, 00:27:57.335 "num_base_bdevs_operational": 3, 00:27:57.335 "base_bdevs_list": [ 00:27:57.335 { 00:27:57.335 "name": null, 00:27:57.335 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:57.335 "is_configured": false, 00:27:57.335 "data_offset": 0, 00:27:57.335 "data_size": 63488 00:27:57.335 }, 00:27:57.335 { 00:27:57.335 "name": "BaseBdev2", 00:27:57.335 "uuid": "b51137bf-830c-5855-b862-e71a62b5d33f", 00:27:57.335 "is_configured": true, 00:27:57.335 "data_offset": 2048, 00:27:57.335 "data_size": 63488 00:27:57.335 }, 00:27:57.335 { 00:27:57.335 "name": "BaseBdev3", 00:27:57.335 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:27:57.335 "is_configured": true, 00:27:57.335 "data_offset": 2048, 00:27:57.335 "data_size": 63488 00:27:57.335 }, 00:27:57.335 { 00:27:57.335 "name": "BaseBdev4", 00:27:57.335 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:27:57.335 "is_configured": true, 00:27:57.335 "data_offset": 2048, 00:27:57.335 "data_size": 63488 00:27:57.335 } 00:27:57.335 ] 00:27:57.335 }' 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:57.335 [2024-10-01 20:26:51.963060] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:27:57.335 [2024-10-01 20:26:51.976789] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca3500 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:57.335 20:26:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:27:57.335 [2024-10-01 20:26:51.979578] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:27:57.902 20:26:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:57.902 20:26:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:57.902 20:26:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:57.902 20:26:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:57.902 20:26:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:57.902 20:26:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:57.902 20:26:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:57.902 20:26:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:57.902 20:26:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:57.902 20:26:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:57.902 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:57.902 "name": "raid_bdev1", 00:27:57.902 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:27:57.902 "strip_size_kb": 0, 00:27:57.902 "state": "online", 00:27:57.902 "raid_level": "raid1", 00:27:57.902 "superblock": true, 00:27:57.902 "num_base_bdevs": 4, 00:27:57.902 "num_base_bdevs_discovered": 4, 00:27:57.902 "num_base_bdevs_operational": 4, 00:27:57.902 "process": { 00:27:57.902 "type": "rebuild", 00:27:57.902 "target": "spare", 00:27:57.902 "progress": { 00:27:57.902 "blocks": 20480, 00:27:57.902 "percent": 32 00:27:57.902 } 00:27:57.902 }, 00:27:57.902 "base_bdevs_list": [ 00:27:57.902 { 00:27:57.902 "name": "spare", 00:27:57.902 "uuid": "775d4c78-8ac5-5db5-8d8a-8ed93921605b", 00:27:57.902 "is_configured": true, 00:27:57.902 "data_offset": 2048, 00:27:57.902 "data_size": 63488 00:27:57.902 }, 00:27:57.902 { 00:27:57.902 "name": "BaseBdev2", 00:27:57.902 "uuid": "b51137bf-830c-5855-b862-e71a62b5d33f", 00:27:57.902 "is_configured": true, 00:27:57.902 "data_offset": 2048, 00:27:57.902 "data_size": 63488 00:27:57.902 }, 00:27:57.902 { 00:27:57.902 "name": "BaseBdev3", 00:27:57.902 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:27:57.902 "is_configured": true, 00:27:57.902 "data_offset": 2048, 00:27:57.902 "data_size": 63488 00:27:57.902 }, 00:27:57.902 { 00:27:57.902 "name": "BaseBdev4", 00:27:57.902 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:27:57.902 "is_configured": true, 00:27:57.902 "data_offset": 2048, 00:27:57.902 "data_size": 63488 00:27:57.902 } 00:27:57.902 ] 00:27:57.902 }' 00:27:57.902 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:57.902 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:57.902 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:57.902 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:27:57.902 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:27:57.902 20:26:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:57.902 20:26:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:57.902 [2024-10-01 20:26:53.152922] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:58.161 [2024-10-01 20:26:53.188917] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:27:58.161 [2024-10-01 20:26:53.189041] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:27:58.161 [2024-10-01 20:26:53.189082] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:27:58.161 [2024-10-01 20:26:53.189100] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:27:58.161 "name": "raid_bdev1", 00:27:58.161 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:27:58.161 "strip_size_kb": 0, 00:27:58.161 "state": "online", 00:27:58.161 "raid_level": "raid1", 00:27:58.161 "superblock": true, 00:27:58.161 "num_base_bdevs": 4, 00:27:58.161 "num_base_bdevs_discovered": 3, 00:27:58.161 "num_base_bdevs_operational": 3, 00:27:58.161 "base_bdevs_list": [ 00:27:58.161 { 00:27:58.161 "name": null, 00:27:58.161 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:58.161 "is_configured": false, 00:27:58.161 "data_offset": 0, 00:27:58.161 "data_size": 63488 00:27:58.161 }, 00:27:58.161 { 00:27:58.161 "name": "BaseBdev2", 00:27:58.161 "uuid": "b51137bf-830c-5855-b862-e71a62b5d33f", 00:27:58.161 "is_configured": true, 00:27:58.161 "data_offset": 2048, 00:27:58.161 "data_size": 63488 00:27:58.161 }, 00:27:58.161 { 00:27:58.161 "name": "BaseBdev3", 00:27:58.161 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:27:58.161 "is_configured": true, 00:27:58.161 "data_offset": 2048, 00:27:58.161 "data_size": 63488 00:27:58.161 }, 00:27:58.161 { 00:27:58.161 "name": "BaseBdev4", 00:27:58.161 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:27:58.161 "is_configured": true, 00:27:58.161 "data_offset": 2048, 00:27:58.161 "data_size": 63488 00:27:58.161 } 00:27:58.161 ] 00:27:58.161 }' 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:27:58.161 20:26:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:58.729 "name": "raid_bdev1", 00:27:58.729 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:27:58.729 "strip_size_kb": 0, 00:27:58.729 "state": "online", 00:27:58.729 "raid_level": "raid1", 00:27:58.729 "superblock": true, 00:27:58.729 "num_base_bdevs": 4, 00:27:58.729 "num_base_bdevs_discovered": 3, 00:27:58.729 "num_base_bdevs_operational": 3, 00:27:58.729 "base_bdevs_list": [ 00:27:58.729 { 00:27:58.729 "name": null, 00:27:58.729 "uuid": "00000000-0000-0000-0000-000000000000", 00:27:58.729 "is_configured": false, 00:27:58.729 "data_offset": 0, 00:27:58.729 "data_size": 63488 00:27:58.729 }, 00:27:58.729 { 00:27:58.729 "name": "BaseBdev2", 00:27:58.729 "uuid": "b51137bf-830c-5855-b862-e71a62b5d33f", 00:27:58.729 "is_configured": true, 00:27:58.729 "data_offset": 2048, 00:27:58.729 "data_size": 63488 00:27:58.729 }, 00:27:58.729 { 00:27:58.729 "name": "BaseBdev3", 00:27:58.729 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:27:58.729 "is_configured": true, 00:27:58.729 "data_offset": 2048, 00:27:58.729 "data_size": 63488 00:27:58.729 }, 00:27:58.729 { 00:27:58.729 "name": "BaseBdev4", 00:27:58.729 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:27:58.729 "is_configured": true, 00:27:58.729 "data_offset": 2048, 00:27:58.729 "data_size": 63488 00:27:58.729 } 00:27:58.729 ] 00:27:58.729 }' 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:58.729 [2024-10-01 20:26:53.887521] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:27:58.729 [2024-10-01 20:26:53.900391] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000ca35d0 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:58.729 20:26:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:27:58.729 [2024-10-01 20:26:53.903237] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:27:59.665 20:26:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:27:59.665 20:26:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:27:59.665 20:26:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:27:59.665 20:26:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:27:59.665 20:26:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:27:59.665 20:26:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:27:59.665 20:26:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:27:59.665 20:26:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:59.665 20:26:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:59.925 20:26:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:59.925 20:26:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:27:59.925 "name": "raid_bdev1", 00:27:59.925 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:27:59.925 "strip_size_kb": 0, 00:27:59.925 "state": "online", 00:27:59.925 "raid_level": "raid1", 00:27:59.925 "superblock": true, 00:27:59.925 "num_base_bdevs": 4, 00:27:59.925 "num_base_bdevs_discovered": 4, 00:27:59.925 "num_base_bdevs_operational": 4, 00:27:59.925 "process": { 00:27:59.926 "type": "rebuild", 00:27:59.926 "target": "spare", 00:27:59.926 "progress": { 00:27:59.926 "blocks": 20480, 00:27:59.926 "percent": 32 00:27:59.926 } 00:27:59.926 }, 00:27:59.926 "base_bdevs_list": [ 00:27:59.926 { 00:27:59.926 "name": "spare", 00:27:59.926 "uuid": "775d4c78-8ac5-5db5-8d8a-8ed93921605b", 00:27:59.926 "is_configured": true, 00:27:59.926 "data_offset": 2048, 00:27:59.926 "data_size": 63488 00:27:59.926 }, 00:27:59.926 { 00:27:59.926 "name": "BaseBdev2", 00:27:59.926 "uuid": "b51137bf-830c-5855-b862-e71a62b5d33f", 00:27:59.926 "is_configured": true, 00:27:59.926 "data_offset": 2048, 00:27:59.926 "data_size": 63488 00:27:59.926 }, 00:27:59.926 { 00:27:59.926 "name": "BaseBdev3", 00:27:59.926 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:27:59.926 "is_configured": true, 00:27:59.926 "data_offset": 2048, 00:27:59.926 "data_size": 63488 00:27:59.926 }, 00:27:59.926 { 00:27:59.926 "name": "BaseBdev4", 00:27:59.926 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:27:59.926 "is_configured": true, 00:27:59.926 "data_offset": 2048, 00:27:59.926 "data_size": 63488 00:27:59.926 } 00:27:59.926 ] 00:27:59.926 }' 00:27:59.926 20:26:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:27:59.926 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:27:59.926 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:27:59.926 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:27:59.926 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:27:59.926 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:27:59.926 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:27:59.926 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:27:59.926 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:27:59.926 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:27:59.926 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:27:59.926 20:26:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:59.926 20:26:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:27:59.926 [2024-10-01 20:26:55.076385] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:28:00.183 [2024-10-01 20:26:55.212549] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000ca35d0 00:28:00.183 20:26:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:00.183 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:28:00.183 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:28:00.183 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:00.184 "name": "raid_bdev1", 00:28:00.184 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:00.184 "strip_size_kb": 0, 00:28:00.184 "state": "online", 00:28:00.184 "raid_level": "raid1", 00:28:00.184 "superblock": true, 00:28:00.184 "num_base_bdevs": 4, 00:28:00.184 "num_base_bdevs_discovered": 3, 00:28:00.184 "num_base_bdevs_operational": 3, 00:28:00.184 "process": { 00:28:00.184 "type": "rebuild", 00:28:00.184 "target": "spare", 00:28:00.184 "progress": { 00:28:00.184 "blocks": 24576, 00:28:00.184 "percent": 38 00:28:00.184 } 00:28:00.184 }, 00:28:00.184 "base_bdevs_list": [ 00:28:00.184 { 00:28:00.184 "name": "spare", 00:28:00.184 "uuid": "775d4c78-8ac5-5db5-8d8a-8ed93921605b", 00:28:00.184 "is_configured": true, 00:28:00.184 "data_offset": 2048, 00:28:00.184 "data_size": 63488 00:28:00.184 }, 00:28:00.184 { 00:28:00.184 "name": null, 00:28:00.184 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:00.184 "is_configured": false, 00:28:00.184 "data_offset": 0, 00:28:00.184 "data_size": 63488 00:28:00.184 }, 00:28:00.184 { 00:28:00.184 "name": "BaseBdev3", 00:28:00.184 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:00.184 "is_configured": true, 00:28:00.184 "data_offset": 2048, 00:28:00.184 "data_size": 63488 00:28:00.184 }, 00:28:00.184 { 00:28:00.184 "name": "BaseBdev4", 00:28:00.184 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:00.184 "is_configured": true, 00:28:00.184 "data_offset": 2048, 00:28:00.184 "data_size": 63488 00:28:00.184 } 00:28:00.184 ] 00:28:00.184 }' 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=554 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:00.184 "name": "raid_bdev1", 00:28:00.184 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:00.184 "strip_size_kb": 0, 00:28:00.184 "state": "online", 00:28:00.184 "raid_level": "raid1", 00:28:00.184 "superblock": true, 00:28:00.184 "num_base_bdevs": 4, 00:28:00.184 "num_base_bdevs_discovered": 3, 00:28:00.184 "num_base_bdevs_operational": 3, 00:28:00.184 "process": { 00:28:00.184 "type": "rebuild", 00:28:00.184 "target": "spare", 00:28:00.184 "progress": { 00:28:00.184 "blocks": 26624, 00:28:00.184 "percent": 41 00:28:00.184 } 00:28:00.184 }, 00:28:00.184 "base_bdevs_list": [ 00:28:00.184 { 00:28:00.184 "name": "spare", 00:28:00.184 "uuid": "775d4c78-8ac5-5db5-8d8a-8ed93921605b", 00:28:00.184 "is_configured": true, 00:28:00.184 "data_offset": 2048, 00:28:00.184 "data_size": 63488 00:28:00.184 }, 00:28:00.184 { 00:28:00.184 "name": null, 00:28:00.184 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:00.184 "is_configured": false, 00:28:00.184 "data_offset": 0, 00:28:00.184 "data_size": 63488 00:28:00.184 }, 00:28:00.184 { 00:28:00.184 "name": "BaseBdev3", 00:28:00.184 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:00.184 "is_configured": true, 00:28:00.184 "data_offset": 2048, 00:28:00.184 "data_size": 63488 00:28:00.184 }, 00:28:00.184 { 00:28:00.184 "name": "BaseBdev4", 00:28:00.184 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:00.184 "is_configured": true, 00:28:00.184 "data_offset": 2048, 00:28:00.184 "data_size": 63488 00:28:00.184 } 00:28:00.184 ] 00:28:00.184 }' 00:28:00.184 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:00.442 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:00.442 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:00.442 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:00.442 20:26:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:28:01.379 20:26:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:28:01.379 20:26:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:01.379 20:26:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:01.379 20:26:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:01.379 20:26:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:01.379 20:26:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:01.379 20:26:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:01.379 20:26:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:01.379 20:26:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:01.379 20:26:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:01.379 20:26:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:01.379 20:26:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:01.379 "name": "raid_bdev1", 00:28:01.379 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:01.379 "strip_size_kb": 0, 00:28:01.379 "state": "online", 00:28:01.379 "raid_level": "raid1", 00:28:01.379 "superblock": true, 00:28:01.379 "num_base_bdevs": 4, 00:28:01.379 "num_base_bdevs_discovered": 3, 00:28:01.379 "num_base_bdevs_operational": 3, 00:28:01.379 "process": { 00:28:01.379 "type": "rebuild", 00:28:01.379 "target": "spare", 00:28:01.379 "progress": { 00:28:01.379 "blocks": 51200, 00:28:01.379 "percent": 80 00:28:01.379 } 00:28:01.379 }, 00:28:01.379 "base_bdevs_list": [ 00:28:01.379 { 00:28:01.379 "name": "spare", 00:28:01.379 "uuid": "775d4c78-8ac5-5db5-8d8a-8ed93921605b", 00:28:01.379 "is_configured": true, 00:28:01.379 "data_offset": 2048, 00:28:01.379 "data_size": 63488 00:28:01.379 }, 00:28:01.379 { 00:28:01.379 "name": null, 00:28:01.379 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:01.379 "is_configured": false, 00:28:01.379 "data_offset": 0, 00:28:01.379 "data_size": 63488 00:28:01.379 }, 00:28:01.379 { 00:28:01.379 "name": "BaseBdev3", 00:28:01.379 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:01.379 "is_configured": true, 00:28:01.379 "data_offset": 2048, 00:28:01.379 "data_size": 63488 00:28:01.379 }, 00:28:01.379 { 00:28:01.379 "name": "BaseBdev4", 00:28:01.379 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:01.379 "is_configured": true, 00:28:01.379 "data_offset": 2048, 00:28:01.379 "data_size": 63488 00:28:01.379 } 00:28:01.379 ] 00:28:01.379 }' 00:28:01.379 20:26:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:01.379 20:26:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:01.379 20:26:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:01.637 20:26:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:01.637 20:26:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:28:01.896 [2024-10-01 20:26:57.127155] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:28:01.896 [2024-10-01 20:26:57.127282] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:28:01.896 [2024-10-01 20:26:57.127482] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:02.464 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:28:02.465 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:02.465 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:02.465 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:02.465 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:02.465 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:02.465 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:02.465 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:02.465 20:26:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:02.465 20:26:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:02.465 20:26:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:02.722 "name": "raid_bdev1", 00:28:02.722 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:02.722 "strip_size_kb": 0, 00:28:02.722 "state": "online", 00:28:02.722 "raid_level": "raid1", 00:28:02.722 "superblock": true, 00:28:02.722 "num_base_bdevs": 4, 00:28:02.722 "num_base_bdevs_discovered": 3, 00:28:02.722 "num_base_bdevs_operational": 3, 00:28:02.722 "base_bdevs_list": [ 00:28:02.722 { 00:28:02.722 "name": "spare", 00:28:02.722 "uuid": "775d4c78-8ac5-5db5-8d8a-8ed93921605b", 00:28:02.722 "is_configured": true, 00:28:02.722 "data_offset": 2048, 00:28:02.722 "data_size": 63488 00:28:02.722 }, 00:28:02.722 { 00:28:02.722 "name": null, 00:28:02.722 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:02.722 "is_configured": false, 00:28:02.722 "data_offset": 0, 00:28:02.722 "data_size": 63488 00:28:02.722 }, 00:28:02.722 { 00:28:02.722 "name": "BaseBdev3", 00:28:02.722 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:02.722 "is_configured": true, 00:28:02.722 "data_offset": 2048, 00:28:02.722 "data_size": 63488 00:28:02.722 }, 00:28:02.722 { 00:28:02.722 "name": "BaseBdev4", 00:28:02.722 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:02.722 "is_configured": true, 00:28:02.722 "data_offset": 2048, 00:28:02.722 "data_size": 63488 00:28:02.722 } 00:28:02.722 ] 00:28:02.722 }' 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:02.722 "name": "raid_bdev1", 00:28:02.722 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:02.722 "strip_size_kb": 0, 00:28:02.722 "state": "online", 00:28:02.722 "raid_level": "raid1", 00:28:02.722 "superblock": true, 00:28:02.722 "num_base_bdevs": 4, 00:28:02.722 "num_base_bdevs_discovered": 3, 00:28:02.722 "num_base_bdevs_operational": 3, 00:28:02.722 "base_bdevs_list": [ 00:28:02.722 { 00:28:02.722 "name": "spare", 00:28:02.722 "uuid": "775d4c78-8ac5-5db5-8d8a-8ed93921605b", 00:28:02.722 "is_configured": true, 00:28:02.722 "data_offset": 2048, 00:28:02.722 "data_size": 63488 00:28:02.722 }, 00:28:02.722 { 00:28:02.722 "name": null, 00:28:02.722 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:02.722 "is_configured": false, 00:28:02.722 "data_offset": 0, 00:28:02.722 "data_size": 63488 00:28:02.722 }, 00:28:02.722 { 00:28:02.722 "name": "BaseBdev3", 00:28:02.722 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:02.722 "is_configured": true, 00:28:02.722 "data_offset": 2048, 00:28:02.722 "data_size": 63488 00:28:02.722 }, 00:28:02.722 { 00:28:02.722 "name": "BaseBdev4", 00:28:02.722 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:02.722 "is_configured": true, 00:28:02.722 "data_offset": 2048, 00:28:02.722 "data_size": 63488 00:28:02.722 } 00:28:02.722 ] 00:28:02.722 }' 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:28:02.722 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:02.982 20:26:57 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:02.982 "name": "raid_bdev1", 00:28:02.982 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:02.982 "strip_size_kb": 0, 00:28:02.982 "state": "online", 00:28:02.982 "raid_level": "raid1", 00:28:02.982 "superblock": true, 00:28:02.982 "num_base_bdevs": 4, 00:28:02.982 "num_base_bdevs_discovered": 3, 00:28:02.982 "num_base_bdevs_operational": 3, 00:28:02.982 "base_bdevs_list": [ 00:28:02.982 { 00:28:02.982 "name": "spare", 00:28:02.982 "uuid": "775d4c78-8ac5-5db5-8d8a-8ed93921605b", 00:28:02.982 "is_configured": true, 00:28:02.982 "data_offset": 2048, 00:28:02.982 "data_size": 63488 00:28:02.982 }, 00:28:02.982 { 00:28:02.982 "name": null, 00:28:02.982 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:02.982 "is_configured": false, 00:28:02.982 "data_offset": 0, 00:28:02.982 "data_size": 63488 00:28:02.982 }, 00:28:02.982 { 00:28:02.982 "name": "BaseBdev3", 00:28:02.982 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:02.982 "is_configured": true, 00:28:02.982 "data_offset": 2048, 00:28:02.982 "data_size": 63488 00:28:02.982 }, 00:28:02.982 { 00:28:02.982 "name": "BaseBdev4", 00:28:02.982 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:02.982 "is_configured": true, 00:28:02.982 "data_offset": 2048, 00:28:02.982 "data_size": 63488 00:28:02.982 } 00:28:02.982 ] 00:28:02.982 }' 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:02.982 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:03.568 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:28:03.568 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:03.568 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:03.568 [2024-10-01 20:26:58.514106] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:03.568 [2024-10-01 20:26:58.514154] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:03.568 [2024-10-01 20:26:58.514272] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:03.568 [2024-10-01 20:26:58.514385] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:03.568 [2024-10-01 20:26:58.514404] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:28:03.568 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:03.568 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:03.568 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:28:03.568 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:03.568 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:03.568 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:03.568 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:28:03.568 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:28:03.568 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:28:03.568 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:28:03.568 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:28:03.568 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:28:03.569 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:28:03.569 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:28:03.569 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:28:03.569 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:28:03.569 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:28:03.569 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:28:03.569 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:28:03.826 /dev/nbd0 00:28:03.826 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:28:03.826 20:26:58 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:28:03.826 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:28:03.826 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:28:03.826 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:28:03.826 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:28:03.826 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:28:03.826 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:28:03.826 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:28:03.826 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:28:03.826 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:28:03.826 1+0 records in 00:28:03.826 1+0 records out 00:28:03.826 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000360931 s, 11.3 MB/s 00:28:03.826 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:28:03.826 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:28:03.826 20:26:58 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:28:03.826 20:26:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:28:03.826 20:26:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:28:03.826 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:28:03.826 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:28:03.826 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:28:04.085 /dev/nbd1 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:28:04.085 1+0 records in 00:28:04.085 1+0 records out 00:28:04.085 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000481358 s, 8.5 MB/s 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:28:04.085 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:28:04.342 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:28:04.342 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:28:04.342 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:28:04.342 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:28:04.342 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:28:04.342 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:28:04.342 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:28:04.601 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:28:04.601 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:28:04.601 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:28:04.601 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:28:04.601 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:28:04.601 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:28:04.601 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:28:04.601 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:28:04.601 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:28:04.601 20:26:59 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:05.167 [2024-10-01 20:27:00.154975] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:28:05.167 [2024-10-01 20:27:00.155056] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:05.167 [2024-10-01 20:27:00.155097] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:28:05.167 [2024-10-01 20:27:00.155115] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:05.167 [2024-10-01 20:27:00.158269] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:05.167 [2024-10-01 20:27:00.158316] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:28:05.167 [2024-10-01 20:27:00.158452] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:28:05.167 [2024-10-01 20:27:00.158521] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:05.167 [2024-10-01 20:27:00.158745] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:05.167 [2024-10-01 20:27:00.158903] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:28:05.167 spare 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:05.167 [2024-10-01 20:27:00.259109] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:28:05.167 [2024-10-01 20:27:00.259168] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:28:05.167 [2024-10-01 20:27:00.259622] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1c80 00:28:05.167 [2024-10-01 20:27:00.259915] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:28:05.167 [2024-10-01 20:27:00.259938] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:28:05.167 [2024-10-01 20:27:00.260173] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:05.167 "name": "raid_bdev1", 00:28:05.167 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:05.167 "strip_size_kb": 0, 00:28:05.167 "state": "online", 00:28:05.167 "raid_level": "raid1", 00:28:05.167 "superblock": true, 00:28:05.167 "num_base_bdevs": 4, 00:28:05.167 "num_base_bdevs_discovered": 3, 00:28:05.167 "num_base_bdevs_operational": 3, 00:28:05.167 "base_bdevs_list": [ 00:28:05.167 { 00:28:05.167 "name": "spare", 00:28:05.167 "uuid": "775d4c78-8ac5-5db5-8d8a-8ed93921605b", 00:28:05.167 "is_configured": true, 00:28:05.167 "data_offset": 2048, 00:28:05.167 "data_size": 63488 00:28:05.167 }, 00:28:05.167 { 00:28:05.167 "name": null, 00:28:05.167 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:05.167 "is_configured": false, 00:28:05.167 "data_offset": 2048, 00:28:05.167 "data_size": 63488 00:28:05.167 }, 00:28:05.167 { 00:28:05.167 "name": "BaseBdev3", 00:28:05.167 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:05.167 "is_configured": true, 00:28:05.167 "data_offset": 2048, 00:28:05.167 "data_size": 63488 00:28:05.167 }, 00:28:05.167 { 00:28:05.167 "name": "BaseBdev4", 00:28:05.167 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:05.167 "is_configured": true, 00:28:05.167 "data_offset": 2048, 00:28:05.167 "data_size": 63488 00:28:05.167 } 00:28:05.167 ] 00:28:05.167 }' 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:05.167 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:05.733 "name": "raid_bdev1", 00:28:05.733 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:05.733 "strip_size_kb": 0, 00:28:05.733 "state": "online", 00:28:05.733 "raid_level": "raid1", 00:28:05.733 "superblock": true, 00:28:05.733 "num_base_bdevs": 4, 00:28:05.733 "num_base_bdevs_discovered": 3, 00:28:05.733 "num_base_bdevs_operational": 3, 00:28:05.733 "base_bdevs_list": [ 00:28:05.733 { 00:28:05.733 "name": "spare", 00:28:05.733 "uuid": "775d4c78-8ac5-5db5-8d8a-8ed93921605b", 00:28:05.733 "is_configured": true, 00:28:05.733 "data_offset": 2048, 00:28:05.733 "data_size": 63488 00:28:05.733 }, 00:28:05.733 { 00:28:05.733 "name": null, 00:28:05.733 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:05.733 "is_configured": false, 00:28:05.733 "data_offset": 2048, 00:28:05.733 "data_size": 63488 00:28:05.733 }, 00:28:05.733 { 00:28:05.733 "name": "BaseBdev3", 00:28:05.733 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:05.733 "is_configured": true, 00:28:05.733 "data_offset": 2048, 00:28:05.733 "data_size": 63488 00:28:05.733 }, 00:28:05.733 { 00:28:05.733 "name": "BaseBdev4", 00:28:05.733 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:05.733 "is_configured": true, 00:28:05.733 "data_offset": 2048, 00:28:05.733 "data_size": 63488 00:28:05.733 } 00:28:05.733 ] 00:28:05.733 }' 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:05.733 [2024-10-01 20:27:00.955332] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:05.733 20:27:00 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:05.991 20:27:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:05.991 "name": "raid_bdev1", 00:28:05.991 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:05.991 "strip_size_kb": 0, 00:28:05.991 "state": "online", 00:28:05.991 "raid_level": "raid1", 00:28:05.991 "superblock": true, 00:28:05.991 "num_base_bdevs": 4, 00:28:05.991 "num_base_bdevs_discovered": 2, 00:28:05.991 "num_base_bdevs_operational": 2, 00:28:05.991 "base_bdevs_list": [ 00:28:05.991 { 00:28:05.991 "name": null, 00:28:05.991 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:05.991 "is_configured": false, 00:28:05.991 "data_offset": 0, 00:28:05.991 "data_size": 63488 00:28:05.991 }, 00:28:05.991 { 00:28:05.991 "name": null, 00:28:05.991 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:05.991 "is_configured": false, 00:28:05.991 "data_offset": 2048, 00:28:05.991 "data_size": 63488 00:28:05.991 }, 00:28:05.991 { 00:28:05.991 "name": "BaseBdev3", 00:28:05.991 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:05.991 "is_configured": true, 00:28:05.991 "data_offset": 2048, 00:28:05.991 "data_size": 63488 00:28:05.991 }, 00:28:05.991 { 00:28:05.991 "name": "BaseBdev4", 00:28:05.991 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:05.991 "is_configured": true, 00:28:05.991 "data_offset": 2048, 00:28:05.991 "data_size": 63488 00:28:05.991 } 00:28:05.991 ] 00:28:05.991 }' 00:28:05.991 20:27:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:05.991 20:27:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:06.249 20:27:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:28:06.249 20:27:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:06.249 20:27:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:06.249 [2024-10-01 20:27:01.471472] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:06.249 [2024-10-01 20:27:01.471755] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:28:06.249 [2024-10-01 20:27:01.471778] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:28:06.249 [2024-10-01 20:27:01.471832] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:06.249 [2024-10-01 20:27:01.484672] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1d50 00:28:06.249 20:27:01 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:06.249 20:27:01 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:28:06.249 [2024-10-01 20:27:01.487529] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:07.628 "name": "raid_bdev1", 00:28:07.628 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:07.628 "strip_size_kb": 0, 00:28:07.628 "state": "online", 00:28:07.628 "raid_level": "raid1", 00:28:07.628 "superblock": true, 00:28:07.628 "num_base_bdevs": 4, 00:28:07.628 "num_base_bdevs_discovered": 3, 00:28:07.628 "num_base_bdevs_operational": 3, 00:28:07.628 "process": { 00:28:07.628 "type": "rebuild", 00:28:07.628 "target": "spare", 00:28:07.628 "progress": { 00:28:07.628 "blocks": 20480, 00:28:07.628 "percent": 32 00:28:07.628 } 00:28:07.628 }, 00:28:07.628 "base_bdevs_list": [ 00:28:07.628 { 00:28:07.628 "name": "spare", 00:28:07.628 "uuid": "775d4c78-8ac5-5db5-8d8a-8ed93921605b", 00:28:07.628 "is_configured": true, 00:28:07.628 "data_offset": 2048, 00:28:07.628 "data_size": 63488 00:28:07.628 }, 00:28:07.628 { 00:28:07.628 "name": null, 00:28:07.628 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:07.628 "is_configured": false, 00:28:07.628 "data_offset": 2048, 00:28:07.628 "data_size": 63488 00:28:07.628 }, 00:28:07.628 { 00:28:07.628 "name": "BaseBdev3", 00:28:07.628 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:07.628 "is_configured": true, 00:28:07.628 "data_offset": 2048, 00:28:07.628 "data_size": 63488 00:28:07.628 }, 00:28:07.628 { 00:28:07.628 "name": "BaseBdev4", 00:28:07.628 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:07.628 "is_configured": true, 00:28:07.628 "data_offset": 2048, 00:28:07.628 "data_size": 63488 00:28:07.628 } 00:28:07.628 ] 00:28:07.628 }' 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.628 [2024-10-01 20:27:02.652952] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:07.628 [2024-10-01 20:27:02.696886] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:28:07.628 [2024-10-01 20:27:02.697001] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:07.628 [2024-10-01 20:27:02.697034] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:07.628 [2024-10-01 20:27:02.697047] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:07.628 "name": "raid_bdev1", 00:28:07.628 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:07.628 "strip_size_kb": 0, 00:28:07.628 "state": "online", 00:28:07.628 "raid_level": "raid1", 00:28:07.628 "superblock": true, 00:28:07.628 "num_base_bdevs": 4, 00:28:07.628 "num_base_bdevs_discovered": 2, 00:28:07.628 "num_base_bdevs_operational": 2, 00:28:07.628 "base_bdevs_list": [ 00:28:07.628 { 00:28:07.628 "name": null, 00:28:07.628 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:07.628 "is_configured": false, 00:28:07.628 "data_offset": 0, 00:28:07.628 "data_size": 63488 00:28:07.628 }, 00:28:07.628 { 00:28:07.628 "name": null, 00:28:07.628 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:07.628 "is_configured": false, 00:28:07.628 "data_offset": 2048, 00:28:07.628 "data_size": 63488 00:28:07.628 }, 00:28:07.628 { 00:28:07.628 "name": "BaseBdev3", 00:28:07.628 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:07.628 "is_configured": true, 00:28:07.628 "data_offset": 2048, 00:28:07.628 "data_size": 63488 00:28:07.628 }, 00:28:07.628 { 00:28:07.628 "name": "BaseBdev4", 00:28:07.628 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:07.628 "is_configured": true, 00:28:07.628 "data_offset": 2048, 00:28:07.628 "data_size": 63488 00:28:07.628 } 00:28:07.628 ] 00:28:07.628 }' 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:07.628 20:27:02 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:08.195 20:27:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:28:08.195 20:27:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:08.195 20:27:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:08.195 [2024-10-01 20:27:03.235555] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:28:08.195 [2024-10-01 20:27:03.235660] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:08.195 [2024-10-01 20:27:03.235745] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c380 00:28:08.195 [2024-10-01 20:27:03.235766] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:08.195 [2024-10-01 20:27:03.236458] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:08.195 [2024-10-01 20:27:03.236493] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:28:08.195 [2024-10-01 20:27:03.236627] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:28:08.195 [2024-10-01 20:27:03.236649] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:28:08.195 [2024-10-01 20:27:03.236667] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:28:08.195 [2024-10-01 20:27:03.236700] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:08.195 [2024-10-01 20:27:03.249457] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cc1e20 00:28:08.195 spare 00:28:08.195 20:27:03 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:08.195 20:27:03 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:28:08.195 [2024-10-01 20:27:03.252247] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:28:09.131 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:09.131 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:09.131 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:09.131 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:09.131 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:09.131 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:09.131 20:27:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:09.131 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:09.131 20:27:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.131 20:27:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:09.131 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:09.131 "name": "raid_bdev1", 00:28:09.131 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:09.131 "strip_size_kb": 0, 00:28:09.131 "state": "online", 00:28:09.131 "raid_level": "raid1", 00:28:09.131 "superblock": true, 00:28:09.131 "num_base_bdevs": 4, 00:28:09.131 "num_base_bdevs_discovered": 3, 00:28:09.131 "num_base_bdevs_operational": 3, 00:28:09.131 "process": { 00:28:09.131 "type": "rebuild", 00:28:09.131 "target": "spare", 00:28:09.131 "progress": { 00:28:09.131 "blocks": 20480, 00:28:09.131 "percent": 32 00:28:09.131 } 00:28:09.131 }, 00:28:09.131 "base_bdevs_list": [ 00:28:09.131 { 00:28:09.131 "name": "spare", 00:28:09.131 "uuid": "775d4c78-8ac5-5db5-8d8a-8ed93921605b", 00:28:09.131 "is_configured": true, 00:28:09.131 "data_offset": 2048, 00:28:09.131 "data_size": 63488 00:28:09.131 }, 00:28:09.131 { 00:28:09.131 "name": null, 00:28:09.131 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:09.131 "is_configured": false, 00:28:09.131 "data_offset": 2048, 00:28:09.131 "data_size": 63488 00:28:09.131 }, 00:28:09.131 { 00:28:09.131 "name": "BaseBdev3", 00:28:09.131 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:09.131 "is_configured": true, 00:28:09.131 "data_offset": 2048, 00:28:09.131 "data_size": 63488 00:28:09.131 }, 00:28:09.131 { 00:28:09.131 "name": "BaseBdev4", 00:28:09.131 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:09.131 "is_configured": true, 00:28:09.131 "data_offset": 2048, 00:28:09.131 "data_size": 63488 00:28:09.131 } 00:28:09.131 ] 00:28:09.131 }' 00:28:09.131 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:09.131 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:09.131 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.390 [2024-10-01 20:27:04.401551] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:09.390 [2024-10-01 20:27:04.461554] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:28:09.390 [2024-10-01 20:27:04.461684] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:09.390 [2024-10-01 20:27:04.461741] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:09.390 [2024-10-01 20:27:04.461762] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:09.390 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:09.390 "name": "raid_bdev1", 00:28:09.390 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:09.390 "strip_size_kb": 0, 00:28:09.390 "state": "online", 00:28:09.390 "raid_level": "raid1", 00:28:09.390 "superblock": true, 00:28:09.390 "num_base_bdevs": 4, 00:28:09.390 "num_base_bdevs_discovered": 2, 00:28:09.390 "num_base_bdevs_operational": 2, 00:28:09.390 "base_bdevs_list": [ 00:28:09.390 { 00:28:09.390 "name": null, 00:28:09.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:09.390 "is_configured": false, 00:28:09.390 "data_offset": 0, 00:28:09.390 "data_size": 63488 00:28:09.390 }, 00:28:09.390 { 00:28:09.390 "name": null, 00:28:09.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:09.390 "is_configured": false, 00:28:09.391 "data_offset": 2048, 00:28:09.391 "data_size": 63488 00:28:09.391 }, 00:28:09.391 { 00:28:09.391 "name": "BaseBdev3", 00:28:09.391 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:09.391 "is_configured": true, 00:28:09.391 "data_offset": 2048, 00:28:09.391 "data_size": 63488 00:28:09.391 }, 00:28:09.391 { 00:28:09.391 "name": "BaseBdev4", 00:28:09.391 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:09.391 "is_configured": true, 00:28:09.391 "data_offset": 2048, 00:28:09.391 "data_size": 63488 00:28:09.391 } 00:28:09.391 ] 00:28:09.391 }' 00:28:09.391 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:09.391 20:27:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.958 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:09.958 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:09.959 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:28:09.959 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:28:09.959 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:09.959 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:09.959 20:27:04 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:09.959 20:27:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:09.959 20:27:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.959 20:27:04 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:09.959 20:27:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:09.959 "name": "raid_bdev1", 00:28:09.959 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:09.959 "strip_size_kb": 0, 00:28:09.959 "state": "online", 00:28:09.959 "raid_level": "raid1", 00:28:09.959 "superblock": true, 00:28:09.959 "num_base_bdevs": 4, 00:28:09.959 "num_base_bdevs_discovered": 2, 00:28:09.959 "num_base_bdevs_operational": 2, 00:28:09.959 "base_bdevs_list": [ 00:28:09.959 { 00:28:09.959 "name": null, 00:28:09.959 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:09.959 "is_configured": false, 00:28:09.959 "data_offset": 0, 00:28:09.959 "data_size": 63488 00:28:09.959 }, 00:28:09.959 { 00:28:09.959 "name": null, 00:28:09.959 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:09.959 "is_configured": false, 00:28:09.959 "data_offset": 2048, 00:28:09.959 "data_size": 63488 00:28:09.959 }, 00:28:09.959 { 00:28:09.959 "name": "BaseBdev3", 00:28:09.959 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:09.959 "is_configured": true, 00:28:09.959 "data_offset": 2048, 00:28:09.959 "data_size": 63488 00:28:09.959 }, 00:28:09.959 { 00:28:09.959 "name": "BaseBdev4", 00:28:09.959 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:09.959 "is_configured": true, 00:28:09.959 "data_offset": 2048, 00:28:09.959 "data_size": 63488 00:28:09.959 } 00:28:09.959 ] 00:28:09.959 }' 00:28:09.959 20:27:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:09.959 20:27:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:28:09.959 20:27:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:09.959 20:27:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:28:09.959 20:27:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:28:09.959 20:27:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:09.959 20:27:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.959 20:27:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:09.959 20:27:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:28:09.959 20:27:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:09.959 20:27:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:09.959 [2024-10-01 20:27:05.140211] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:28:09.959 [2024-10-01 20:27:05.140520] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:09.959 [2024-10-01 20:27:05.140571] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c980 00:28:09.959 [2024-10-01 20:27:05.140598] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:09.959 [2024-10-01 20:27:05.141250] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:09.959 [2024-10-01 20:27:05.141283] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:28:09.959 [2024-10-01 20:27:05.141426] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:28:09.959 [2024-10-01 20:27:05.141456] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:28:09.959 [2024-10-01 20:27:05.141470] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:28:09.959 [2024-10-01 20:27:05.141493] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:28:09.959 BaseBdev1 00:28:09.959 20:27:05 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:09.959 20:27:05 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:28:10.909 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:10.909 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:10.909 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:10.909 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:10.909 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:10.909 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:10.909 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:10.909 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:10.909 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:10.909 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:10.909 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:10.909 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:10.909 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:10.909 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:11.168 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:11.168 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:11.168 "name": "raid_bdev1", 00:28:11.168 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:11.168 "strip_size_kb": 0, 00:28:11.168 "state": "online", 00:28:11.168 "raid_level": "raid1", 00:28:11.168 "superblock": true, 00:28:11.168 "num_base_bdevs": 4, 00:28:11.168 "num_base_bdevs_discovered": 2, 00:28:11.168 "num_base_bdevs_operational": 2, 00:28:11.168 "base_bdevs_list": [ 00:28:11.168 { 00:28:11.168 "name": null, 00:28:11.168 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:11.168 "is_configured": false, 00:28:11.168 "data_offset": 0, 00:28:11.168 "data_size": 63488 00:28:11.168 }, 00:28:11.168 { 00:28:11.168 "name": null, 00:28:11.168 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:11.168 "is_configured": false, 00:28:11.168 "data_offset": 2048, 00:28:11.168 "data_size": 63488 00:28:11.168 }, 00:28:11.168 { 00:28:11.168 "name": "BaseBdev3", 00:28:11.168 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:11.168 "is_configured": true, 00:28:11.168 "data_offset": 2048, 00:28:11.168 "data_size": 63488 00:28:11.168 }, 00:28:11.168 { 00:28:11.168 "name": "BaseBdev4", 00:28:11.168 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:11.168 "is_configured": true, 00:28:11.168 "data_offset": 2048, 00:28:11.168 "data_size": 63488 00:28:11.168 } 00:28:11.168 ] 00:28:11.168 }' 00:28:11.168 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:11.168 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:11.427 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:11.427 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:11.427 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:28:11.427 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:28:11.427 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:11.427 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:11.427 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:11.427 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:11.427 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:11.427 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:11.686 "name": "raid_bdev1", 00:28:11.686 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:11.686 "strip_size_kb": 0, 00:28:11.686 "state": "online", 00:28:11.686 "raid_level": "raid1", 00:28:11.686 "superblock": true, 00:28:11.686 "num_base_bdevs": 4, 00:28:11.686 "num_base_bdevs_discovered": 2, 00:28:11.686 "num_base_bdevs_operational": 2, 00:28:11.686 "base_bdevs_list": [ 00:28:11.686 { 00:28:11.686 "name": null, 00:28:11.686 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:11.686 "is_configured": false, 00:28:11.686 "data_offset": 0, 00:28:11.686 "data_size": 63488 00:28:11.686 }, 00:28:11.686 { 00:28:11.686 "name": null, 00:28:11.686 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:11.686 "is_configured": false, 00:28:11.686 "data_offset": 2048, 00:28:11.686 "data_size": 63488 00:28:11.686 }, 00:28:11.686 { 00:28:11.686 "name": "BaseBdev3", 00:28:11.686 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:11.686 "is_configured": true, 00:28:11.686 "data_offset": 2048, 00:28:11.686 "data_size": 63488 00:28:11.686 }, 00:28:11.686 { 00:28:11.686 "name": "BaseBdev4", 00:28:11.686 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:11.686 "is_configured": true, 00:28:11.686 "data_offset": 2048, 00:28:11.686 "data_size": 63488 00:28:11.686 } 00:28:11.686 ] 00:28:11.686 }' 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:11.686 [2024-10-01 20:27:06.788581] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:11.686 [2024-10-01 20:27:06.788906] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:28:11.686 [2024-10-01 20:27:06.788930] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:28:11.686 request: 00:28:11.686 { 00:28:11.686 "base_bdev": "BaseBdev1", 00:28:11.686 "raid_bdev": "raid_bdev1", 00:28:11.686 "method": "bdev_raid_add_base_bdev", 00:28:11.686 "req_id": 1 00:28:11.686 } 00:28:11.686 Got JSON-RPC error response 00:28:11.686 response: 00:28:11.686 { 00:28:11.686 "code": -22, 00:28:11.686 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:28:11.686 } 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:28:11.686 20:27:06 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:28:12.623 20:27:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:12.623 20:27:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:12.623 20:27:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:12.623 20:27:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:12.623 20:27:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:12.623 20:27:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:12.623 20:27:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:12.623 20:27:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:12.623 20:27:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:12.623 20:27:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:12.623 20:27:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:12.623 20:27:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:12.623 20:27:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:12.623 20:27:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:12.623 20:27:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:12.623 20:27:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:12.623 "name": "raid_bdev1", 00:28:12.624 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:12.624 "strip_size_kb": 0, 00:28:12.624 "state": "online", 00:28:12.624 "raid_level": "raid1", 00:28:12.624 "superblock": true, 00:28:12.624 "num_base_bdevs": 4, 00:28:12.624 "num_base_bdevs_discovered": 2, 00:28:12.624 "num_base_bdevs_operational": 2, 00:28:12.624 "base_bdevs_list": [ 00:28:12.624 { 00:28:12.624 "name": null, 00:28:12.624 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:12.624 "is_configured": false, 00:28:12.624 "data_offset": 0, 00:28:12.624 "data_size": 63488 00:28:12.624 }, 00:28:12.624 { 00:28:12.624 "name": null, 00:28:12.624 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:12.624 "is_configured": false, 00:28:12.624 "data_offset": 2048, 00:28:12.624 "data_size": 63488 00:28:12.624 }, 00:28:12.624 { 00:28:12.624 "name": "BaseBdev3", 00:28:12.624 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:12.624 "is_configured": true, 00:28:12.624 "data_offset": 2048, 00:28:12.624 "data_size": 63488 00:28:12.624 }, 00:28:12.624 { 00:28:12.624 "name": "BaseBdev4", 00:28:12.624 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:12.624 "is_configured": true, 00:28:12.624 "data_offset": 2048, 00:28:12.624 "data_size": 63488 00:28:12.624 } 00:28:12.624 ] 00:28:12.624 }' 00:28:12.624 20:27:07 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:12.624 20:27:07 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:13.190 20:27:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:13.190 20:27:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:13.190 20:27:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:28:13.190 20:27:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:28:13.190 20:27:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:13.190 20:27:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:13.190 20:27:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:13.190 20:27:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:13.190 20:27:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:13.190 20:27:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:13.190 20:27:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:13.190 "name": "raid_bdev1", 00:28:13.190 "uuid": "92be3651-baab-44a1-92c6-299a3817fa30", 00:28:13.190 "strip_size_kb": 0, 00:28:13.190 "state": "online", 00:28:13.190 "raid_level": "raid1", 00:28:13.190 "superblock": true, 00:28:13.190 "num_base_bdevs": 4, 00:28:13.190 "num_base_bdevs_discovered": 2, 00:28:13.190 "num_base_bdevs_operational": 2, 00:28:13.190 "base_bdevs_list": [ 00:28:13.190 { 00:28:13.190 "name": null, 00:28:13.190 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:13.190 "is_configured": false, 00:28:13.190 "data_offset": 0, 00:28:13.190 "data_size": 63488 00:28:13.190 }, 00:28:13.190 { 00:28:13.190 "name": null, 00:28:13.190 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:13.190 "is_configured": false, 00:28:13.190 "data_offset": 2048, 00:28:13.190 "data_size": 63488 00:28:13.190 }, 00:28:13.190 { 00:28:13.190 "name": "BaseBdev3", 00:28:13.190 "uuid": "65da92b6-6013-5043-9652-e8d45074d7c4", 00:28:13.190 "is_configured": true, 00:28:13.190 "data_offset": 2048, 00:28:13.190 "data_size": 63488 00:28:13.190 }, 00:28:13.190 { 00:28:13.190 "name": "BaseBdev4", 00:28:13.190 "uuid": "e37121aa-edbc-5447-86e9-47ad3ebfa660", 00:28:13.190 "is_configured": true, 00:28:13.190 "data_offset": 2048, 00:28:13.190 "data_size": 63488 00:28:13.190 } 00:28:13.190 ] 00:28:13.190 }' 00:28:13.190 20:27:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:13.190 20:27:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:28:13.190 20:27:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:13.449 20:27:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:28:13.449 20:27:08 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 79094 00:28:13.449 20:27:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@950 -- # '[' -z 79094 ']' 00:28:13.449 20:27:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # kill -0 79094 00:28:13.449 20:27:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@955 -- # uname 00:28:13.449 20:27:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:28:13.449 20:27:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 79094 00:28:13.449 killing process with pid 79094 00:28:13.449 Received shutdown signal, test time was about 60.000000 seconds 00:28:13.449 00:28:13.449 Latency(us) 00:28:13.449 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:13.449 =================================================================================================================== 00:28:13.449 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:28:13.449 20:27:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:28:13.449 20:27:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:28:13.449 20:27:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 79094' 00:28:13.449 20:27:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@969 -- # kill 79094 00:28:13.449 [2024-10-01 20:27:08.508705] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:28:13.449 20:27:08 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@974 -- # wait 79094 00:28:13.449 [2024-10-01 20:27:08.508933] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:13.449 [2024-10-01 20:27:08.509036] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:13.449 [2024-10-01 20:27:08.509056] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:28:14.016 [2024-10-01 20:27:08.965148] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:28:15.923 ************************************ 00:28:15.923 END TEST raid_rebuild_test_sb 00:28:15.923 ************************************ 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:28:15.923 00:28:15.923 real 0m30.686s 00:28:15.923 user 0m36.250s 00:28:15.923 sys 0m4.359s 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:28:15.923 20:27:10 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 4 false true true 00:28:15.923 20:27:10 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:28:15.923 20:27:10 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:28:15.923 20:27:10 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:28:15.923 ************************************ 00:28:15.923 START TEST raid_rebuild_test_io 00:28:15.923 ************************************ 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 4 false true true 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=79904 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 79904 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@831 -- # '[' -z 79904 ']' 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@836 -- # local max_retries=100 00:28:15.923 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # xtrace_disable 00:28:15.923 20:27:10 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:15.923 I/O size of 3145728 is greater than zero copy threshold (65536). 00:28:15.923 Zero copy mechanism will not be used. 00:28:15.923 [2024-10-01 20:27:10.957819] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:28:15.923 [2024-10-01 20:27:10.958021] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79904 ] 00:28:15.923 [2024-10-01 20:27:11.138038] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:16.182 [2024-10-01 20:27:11.384798] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:28:16.441 [2024-10-01 20:27:11.592409] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:16.441 [2024-10-01 20:27:11.592794] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # return 0 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.009 BaseBdev1_malloc 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.009 [2024-10-01 20:27:12.060680] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:28:17.009 [2024-10-01 20:27:12.060942] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:17.009 [2024-10-01 20:27:12.061029] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:28:17.009 [2024-10-01 20:27:12.061324] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:17.009 [2024-10-01 20:27:12.064555] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:17.009 [2024-10-01 20:27:12.064735] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:28:17.009 BaseBdev1 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.009 BaseBdev2_malloc 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.009 [2024-10-01 20:27:12.118370] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:28:17.009 [2024-10-01 20:27:12.118602] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:17.009 [2024-10-01 20:27:12.118651] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:28:17.009 [2024-10-01 20:27:12.118676] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:17.009 [2024-10-01 20:27:12.121661] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:17.009 [2024-10-01 20:27:12.121864] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:28:17.009 BaseBdev2 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.009 BaseBdev3_malloc 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.009 [2024-10-01 20:27:12.175075] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:28:17.009 [2024-10-01 20:27:12.175158] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:17.009 [2024-10-01 20:27:12.175194] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:28:17.009 [2024-10-01 20:27:12.175213] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:17.009 [2024-10-01 20:27:12.178264] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:17.009 [2024-10-01 20:27:12.178318] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:28:17.009 BaseBdev3 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.009 BaseBdev4_malloc 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.009 [2024-10-01 20:27:12.228258] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:28:17.009 [2024-10-01 20:27:12.228338] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:17.009 [2024-10-01 20:27:12.228370] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:28:17.009 [2024-10-01 20:27:12.228389] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:17.009 [2024-10-01 20:27:12.231431] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:17.009 [2024-10-01 20:27:12.231486] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:28:17.009 BaseBdev4 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.009 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.268 spare_malloc 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.268 spare_delay 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.268 [2024-10-01 20:27:12.297171] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:28:17.268 [2024-10-01 20:27:12.297259] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:17.268 [2024-10-01 20:27:12.297316] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:28:17.268 [2024-10-01 20:27:12.297339] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:17.268 [2024-10-01 20:27:12.300345] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:17.268 [2024-10-01 20:27:12.300405] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:28:17.268 spare 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.268 [2024-10-01 20:27:12.309434] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:17.268 [2024-10-01 20:27:12.312030] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:17.268 [2024-10-01 20:27:12.312142] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:17.268 [2024-10-01 20:27:12.312229] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:28:17.268 [2024-10-01 20:27:12.312364] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:28:17.268 [2024-10-01 20:27:12.312384] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:28:17.268 [2024-10-01 20:27:12.312783] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:28:17.268 [2024-10-01 20:27:12.313021] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:28:17.268 [2024-10-01 20:27:12.313040] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:28:17.268 [2024-10-01 20:27:12.313325] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.268 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:17.268 "name": "raid_bdev1", 00:28:17.268 "uuid": "a0828108-1a63-4e8f-9294-52cbf3f69ca9", 00:28:17.268 "strip_size_kb": 0, 00:28:17.268 "state": "online", 00:28:17.268 "raid_level": "raid1", 00:28:17.268 "superblock": false, 00:28:17.268 "num_base_bdevs": 4, 00:28:17.268 "num_base_bdevs_discovered": 4, 00:28:17.268 "num_base_bdevs_operational": 4, 00:28:17.268 "base_bdevs_list": [ 00:28:17.268 { 00:28:17.269 "name": "BaseBdev1", 00:28:17.269 "uuid": "fada2cf1-c889-5144-87cf-98148165362f", 00:28:17.269 "is_configured": true, 00:28:17.269 "data_offset": 0, 00:28:17.269 "data_size": 65536 00:28:17.269 }, 00:28:17.269 { 00:28:17.269 "name": "BaseBdev2", 00:28:17.269 "uuid": "50d776db-d85e-58d2-b7de-d64b8fd032df", 00:28:17.269 "is_configured": true, 00:28:17.269 "data_offset": 0, 00:28:17.269 "data_size": 65536 00:28:17.269 }, 00:28:17.269 { 00:28:17.269 "name": "BaseBdev3", 00:28:17.269 "uuid": "b46da91b-a0df-5787-9186-77651356325b", 00:28:17.269 "is_configured": true, 00:28:17.269 "data_offset": 0, 00:28:17.269 "data_size": 65536 00:28:17.269 }, 00:28:17.269 { 00:28:17.269 "name": "BaseBdev4", 00:28:17.269 "uuid": "66a5765f-21f4-5295-a031-7044d144ce73", 00:28:17.269 "is_configured": true, 00:28:17.269 "data_offset": 0, 00:28:17.269 "data_size": 65536 00:28:17.269 } 00:28:17.269 ] 00:28:17.269 }' 00:28:17.269 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:17.269 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.623 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:28:17.623 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.623 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:28:17.623 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.623 [2024-10-01 20:27:12.829981] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:17.623 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.623 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.911 [2024-10-01 20:27:12.913544] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:17.911 "name": "raid_bdev1", 00:28:17.911 "uuid": "a0828108-1a63-4e8f-9294-52cbf3f69ca9", 00:28:17.911 "strip_size_kb": 0, 00:28:17.911 "state": "online", 00:28:17.911 "raid_level": "raid1", 00:28:17.911 "superblock": false, 00:28:17.911 "num_base_bdevs": 4, 00:28:17.911 "num_base_bdevs_discovered": 3, 00:28:17.911 "num_base_bdevs_operational": 3, 00:28:17.911 "base_bdevs_list": [ 00:28:17.911 { 00:28:17.911 "name": null, 00:28:17.911 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:17.911 "is_configured": false, 00:28:17.911 "data_offset": 0, 00:28:17.911 "data_size": 65536 00:28:17.911 }, 00:28:17.911 { 00:28:17.911 "name": "BaseBdev2", 00:28:17.911 "uuid": "50d776db-d85e-58d2-b7de-d64b8fd032df", 00:28:17.911 "is_configured": true, 00:28:17.911 "data_offset": 0, 00:28:17.911 "data_size": 65536 00:28:17.911 }, 00:28:17.911 { 00:28:17.911 "name": "BaseBdev3", 00:28:17.911 "uuid": "b46da91b-a0df-5787-9186-77651356325b", 00:28:17.911 "is_configured": true, 00:28:17.911 "data_offset": 0, 00:28:17.911 "data_size": 65536 00:28:17.911 }, 00:28:17.911 { 00:28:17.911 "name": "BaseBdev4", 00:28:17.911 "uuid": "66a5765f-21f4-5295-a031-7044d144ce73", 00:28:17.911 "is_configured": true, 00:28:17.911 "data_offset": 0, 00:28:17.911 "data_size": 65536 00:28:17.911 } 00:28:17.911 ] 00:28:17.911 }' 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:17.911 20:27:12 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:17.911 [2024-10-01 20:27:13.017725] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:28:17.911 I/O size of 3145728 is greater than zero copy threshold (65536). 00:28:17.911 Zero copy mechanism will not be used. 00:28:17.911 Running I/O for 60 seconds... 00:28:18.476 20:27:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:28:18.476 20:27:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:18.476 20:27:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:18.477 [2024-10-01 20:27:13.449238] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:18.477 20:27:13 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:18.477 20:27:13 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:28:18.477 [2024-10-01 20:27:13.514760] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:28:18.477 [2024-10-01 20:27:13.517624] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:28:18.477 [2024-10-01 20:27:13.638365] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:28:18.477 [2024-10-01 20:27:13.639228] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:28:18.734 [2024-10-01 20:27:13.866256] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:28:18.993 161.00 IOPS, 483.00 MiB/s [2024-10-01 20:27:14.244479] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:28:19.251 [2024-10-01 20:27:14.487255] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:28:19.251 [2024-10-01 20:27:14.487923] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:19.510 "name": "raid_bdev1", 00:28:19.510 "uuid": "a0828108-1a63-4e8f-9294-52cbf3f69ca9", 00:28:19.510 "strip_size_kb": 0, 00:28:19.510 "state": "online", 00:28:19.510 "raid_level": "raid1", 00:28:19.510 "superblock": false, 00:28:19.510 "num_base_bdevs": 4, 00:28:19.510 "num_base_bdevs_discovered": 4, 00:28:19.510 "num_base_bdevs_operational": 4, 00:28:19.510 "process": { 00:28:19.510 "type": "rebuild", 00:28:19.510 "target": "spare", 00:28:19.510 "progress": { 00:28:19.510 "blocks": 10240, 00:28:19.510 "percent": 15 00:28:19.510 } 00:28:19.510 }, 00:28:19.510 "base_bdevs_list": [ 00:28:19.510 { 00:28:19.510 "name": "spare", 00:28:19.510 "uuid": "af5e50a2-7b9f-5a01-ac8d-3a11fb14aab0", 00:28:19.510 "is_configured": true, 00:28:19.510 "data_offset": 0, 00:28:19.510 "data_size": 65536 00:28:19.510 }, 00:28:19.510 { 00:28:19.510 "name": "BaseBdev2", 00:28:19.510 "uuid": "50d776db-d85e-58d2-b7de-d64b8fd032df", 00:28:19.510 "is_configured": true, 00:28:19.510 "data_offset": 0, 00:28:19.510 "data_size": 65536 00:28:19.510 }, 00:28:19.510 { 00:28:19.510 "name": "BaseBdev3", 00:28:19.510 "uuid": "b46da91b-a0df-5787-9186-77651356325b", 00:28:19.510 "is_configured": true, 00:28:19.510 "data_offset": 0, 00:28:19.510 "data_size": 65536 00:28:19.510 }, 00:28:19.510 { 00:28:19.510 "name": "BaseBdev4", 00:28:19.510 "uuid": "66a5765f-21f4-5295-a031-7044d144ce73", 00:28:19.510 "is_configured": true, 00:28:19.510 "data_offset": 0, 00:28:19.510 "data_size": 65536 00:28:19.510 } 00:28:19.510 ] 00:28:19.510 }' 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:19.510 20:27:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:19.510 [2024-10-01 20:27:14.682008] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:19.771 [2024-10-01 20:27:14.824799] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:28:19.771 [2024-10-01 20:27:14.829419] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:19.771 [2024-10-01 20:27:14.829480] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:19.771 [2024-10-01 20:27:14.829508] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:28:19.771 [2024-10-01 20:27:14.871168] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006220 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:19.771 "name": "raid_bdev1", 00:28:19.771 "uuid": "a0828108-1a63-4e8f-9294-52cbf3f69ca9", 00:28:19.771 "strip_size_kb": 0, 00:28:19.771 "state": "online", 00:28:19.771 "raid_level": "raid1", 00:28:19.771 "superblock": false, 00:28:19.771 "num_base_bdevs": 4, 00:28:19.771 "num_base_bdevs_discovered": 3, 00:28:19.771 "num_base_bdevs_operational": 3, 00:28:19.771 "base_bdevs_list": [ 00:28:19.771 { 00:28:19.771 "name": null, 00:28:19.771 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:19.771 "is_configured": false, 00:28:19.771 "data_offset": 0, 00:28:19.771 "data_size": 65536 00:28:19.771 }, 00:28:19.771 { 00:28:19.771 "name": "BaseBdev2", 00:28:19.771 "uuid": "50d776db-d85e-58d2-b7de-d64b8fd032df", 00:28:19.771 "is_configured": true, 00:28:19.771 "data_offset": 0, 00:28:19.771 "data_size": 65536 00:28:19.771 }, 00:28:19.771 { 00:28:19.771 "name": "BaseBdev3", 00:28:19.771 "uuid": "b46da91b-a0df-5787-9186-77651356325b", 00:28:19.771 "is_configured": true, 00:28:19.771 "data_offset": 0, 00:28:19.771 "data_size": 65536 00:28:19.771 }, 00:28:19.771 { 00:28:19.771 "name": "BaseBdev4", 00:28:19.771 "uuid": "66a5765f-21f4-5295-a031-7044d144ce73", 00:28:19.771 "is_configured": true, 00:28:19.771 "data_offset": 0, 00:28:19.771 "data_size": 65536 00:28:19.771 } 00:28:19.771 ] 00:28:19.771 }' 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:19.771 20:27:14 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:20.290 132.00 IOPS, 396.00 MiB/s 20:27:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:20.290 20:27:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:20.290 20:27:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:28:20.290 20:27:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:28:20.290 20:27:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:20.290 20:27:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:20.290 20:27:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:20.290 20:27:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:20.290 20:27:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:20.290 20:27:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:20.290 20:27:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:20.290 "name": "raid_bdev1", 00:28:20.290 "uuid": "a0828108-1a63-4e8f-9294-52cbf3f69ca9", 00:28:20.290 "strip_size_kb": 0, 00:28:20.290 "state": "online", 00:28:20.290 "raid_level": "raid1", 00:28:20.290 "superblock": false, 00:28:20.290 "num_base_bdevs": 4, 00:28:20.290 "num_base_bdevs_discovered": 3, 00:28:20.290 "num_base_bdevs_operational": 3, 00:28:20.290 "base_bdevs_list": [ 00:28:20.290 { 00:28:20.290 "name": null, 00:28:20.290 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:20.290 "is_configured": false, 00:28:20.290 "data_offset": 0, 00:28:20.290 "data_size": 65536 00:28:20.290 }, 00:28:20.290 { 00:28:20.290 "name": "BaseBdev2", 00:28:20.290 "uuid": "50d776db-d85e-58d2-b7de-d64b8fd032df", 00:28:20.290 "is_configured": true, 00:28:20.290 "data_offset": 0, 00:28:20.290 "data_size": 65536 00:28:20.290 }, 00:28:20.290 { 00:28:20.290 "name": "BaseBdev3", 00:28:20.290 "uuid": "b46da91b-a0df-5787-9186-77651356325b", 00:28:20.290 "is_configured": true, 00:28:20.290 "data_offset": 0, 00:28:20.290 "data_size": 65536 00:28:20.290 }, 00:28:20.290 { 00:28:20.290 "name": "BaseBdev4", 00:28:20.290 "uuid": "66a5765f-21f4-5295-a031-7044d144ce73", 00:28:20.290 "is_configured": true, 00:28:20.290 "data_offset": 0, 00:28:20.290 "data_size": 65536 00:28:20.290 } 00:28:20.290 ] 00:28:20.290 }' 00:28:20.290 20:27:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:20.549 20:27:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:28:20.549 20:27:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:20.549 20:27:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:28:20.549 20:27:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:28:20.549 20:27:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:20.549 20:27:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:20.549 [2024-10-01 20:27:15.607882] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:20.549 20:27:15 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:20.549 20:27:15 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:28:20.549 [2024-10-01 20:27:15.706530] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:28:20.549 [2024-10-01 20:27:15.709957] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:28:20.807 [2024-10-01 20:27:15.823561] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:28:20.807 [2024-10-01 20:27:15.824334] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:28:20.807 136.00 IOPS, 408.00 MiB/s [2024-10-01 20:27:16.050525] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:28:20.807 [2024-10-01 20:27:16.051622] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:28:21.375 [2024-10-01 20:27:16.424538] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:28:21.375 [2024-10-01 20:27:16.425534] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:28:21.375 [2024-10-01 20:27:16.537216] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:28:21.375 [2024-10-01 20:27:16.538152] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:21.634 "name": "raid_bdev1", 00:28:21.634 "uuid": "a0828108-1a63-4e8f-9294-52cbf3f69ca9", 00:28:21.634 "strip_size_kb": 0, 00:28:21.634 "state": "online", 00:28:21.634 "raid_level": "raid1", 00:28:21.634 "superblock": false, 00:28:21.634 "num_base_bdevs": 4, 00:28:21.634 "num_base_bdevs_discovered": 4, 00:28:21.634 "num_base_bdevs_operational": 4, 00:28:21.634 "process": { 00:28:21.634 "type": "rebuild", 00:28:21.634 "target": "spare", 00:28:21.634 "progress": { 00:28:21.634 "blocks": 10240, 00:28:21.634 "percent": 15 00:28:21.634 } 00:28:21.634 }, 00:28:21.634 "base_bdevs_list": [ 00:28:21.634 { 00:28:21.634 "name": "spare", 00:28:21.634 "uuid": "af5e50a2-7b9f-5a01-ac8d-3a11fb14aab0", 00:28:21.634 "is_configured": true, 00:28:21.634 "data_offset": 0, 00:28:21.634 "data_size": 65536 00:28:21.634 }, 00:28:21.634 { 00:28:21.634 "name": "BaseBdev2", 00:28:21.634 "uuid": "50d776db-d85e-58d2-b7de-d64b8fd032df", 00:28:21.634 "is_configured": true, 00:28:21.634 "data_offset": 0, 00:28:21.634 "data_size": 65536 00:28:21.634 }, 00:28:21.634 { 00:28:21.634 "name": "BaseBdev3", 00:28:21.634 "uuid": "b46da91b-a0df-5787-9186-77651356325b", 00:28:21.634 "is_configured": true, 00:28:21.634 "data_offset": 0, 00:28:21.634 "data_size": 65536 00:28:21.634 }, 00:28:21.634 { 00:28:21.634 "name": "BaseBdev4", 00:28:21.634 "uuid": "66a5765f-21f4-5295-a031-7044d144ce73", 00:28:21.634 "is_configured": true, 00:28:21.634 "data_offset": 0, 00:28:21.634 "data_size": 65536 00:28:21.634 } 00:28:21.634 ] 00:28:21.634 }' 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:21.634 20:27:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:21.634 [2024-10-01 20:27:16.849838] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:28:21.634 [2024-10-01 20:27:16.859202] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:28:21.634 [2024-10-01 20:27:16.877619] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000006220 00:28:21.634 [2024-10-01 20:27:16.877829] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d0000063c0 00:28:21.634 [2024-10-01 20:27:16.877947] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:28:21.893 20:27:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:21.893 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:28:21.893 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:28:21.893 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:21.893 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:21.893 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:21.893 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:21.893 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:21.893 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:21.893 20:27:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:21.893 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:21.893 20:27:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:21.893 20:27:16 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:21.893 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:21.893 "name": "raid_bdev1", 00:28:21.893 "uuid": "a0828108-1a63-4e8f-9294-52cbf3f69ca9", 00:28:21.893 "strip_size_kb": 0, 00:28:21.893 "state": "online", 00:28:21.893 "raid_level": "raid1", 00:28:21.893 "superblock": false, 00:28:21.893 "num_base_bdevs": 4, 00:28:21.893 "num_base_bdevs_discovered": 3, 00:28:21.893 "num_base_bdevs_operational": 3, 00:28:21.893 "process": { 00:28:21.893 "type": "rebuild", 00:28:21.893 "target": "spare", 00:28:21.893 "progress": { 00:28:21.893 "blocks": 14336, 00:28:21.893 "percent": 21 00:28:21.893 } 00:28:21.893 }, 00:28:21.893 "base_bdevs_list": [ 00:28:21.893 { 00:28:21.893 "name": "spare", 00:28:21.893 "uuid": "af5e50a2-7b9f-5a01-ac8d-3a11fb14aab0", 00:28:21.893 "is_configured": true, 00:28:21.893 "data_offset": 0, 00:28:21.893 "data_size": 65536 00:28:21.893 }, 00:28:21.893 { 00:28:21.893 "name": null, 00:28:21.893 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:21.893 "is_configured": false, 00:28:21.893 "data_offset": 0, 00:28:21.893 "data_size": 65536 00:28:21.893 }, 00:28:21.893 { 00:28:21.893 "name": "BaseBdev3", 00:28:21.893 "uuid": "b46da91b-a0df-5787-9186-77651356325b", 00:28:21.893 "is_configured": true, 00:28:21.893 "data_offset": 0, 00:28:21.893 "data_size": 65536 00:28:21.893 }, 00:28:21.893 { 00:28:21.893 "name": "BaseBdev4", 00:28:21.893 "uuid": "66a5765f-21f4-5295-a031-7044d144ce73", 00:28:21.893 "is_configured": true, 00:28:21.893 "data_offset": 0, 00:28:21.893 "data_size": 65536 00:28:21.893 } 00:28:21.893 ] 00:28:21.893 }' 00:28:21.893 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:21.893 20:27:16 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:21.893 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:21.893 120.75 IOPS, 362.25 MiB/s 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:21.893 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=576 00:28:21.893 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:28:21.893 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:21.893 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:21.893 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:21.893 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:21.893 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:21.893 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:21.893 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:21.893 20:27:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:21.893 20:27:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:21.893 20:27:17 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:21.893 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:21.893 "name": "raid_bdev1", 00:28:21.893 "uuid": "a0828108-1a63-4e8f-9294-52cbf3f69ca9", 00:28:21.893 "strip_size_kb": 0, 00:28:21.893 "state": "online", 00:28:21.893 "raid_level": "raid1", 00:28:21.893 "superblock": false, 00:28:21.893 "num_base_bdevs": 4, 00:28:21.893 "num_base_bdevs_discovered": 3, 00:28:21.893 "num_base_bdevs_operational": 3, 00:28:21.894 "process": { 00:28:21.894 "type": "rebuild", 00:28:21.894 "target": "spare", 00:28:21.894 "progress": { 00:28:21.894 "blocks": 14336, 00:28:21.894 "percent": 21 00:28:21.894 } 00:28:21.894 }, 00:28:21.894 "base_bdevs_list": [ 00:28:21.894 { 00:28:21.894 "name": "spare", 00:28:21.894 "uuid": "af5e50a2-7b9f-5a01-ac8d-3a11fb14aab0", 00:28:21.894 "is_configured": true, 00:28:21.894 "data_offset": 0, 00:28:21.894 "data_size": 65536 00:28:21.894 }, 00:28:21.894 { 00:28:21.894 "name": null, 00:28:21.894 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:21.894 "is_configured": false, 00:28:21.894 "data_offset": 0, 00:28:21.894 "data_size": 65536 00:28:21.894 }, 00:28:21.894 { 00:28:21.894 "name": "BaseBdev3", 00:28:21.894 "uuid": "b46da91b-a0df-5787-9186-77651356325b", 00:28:21.894 "is_configured": true, 00:28:21.894 "data_offset": 0, 00:28:21.894 "data_size": 65536 00:28:21.894 }, 00:28:21.894 { 00:28:21.894 "name": "BaseBdev4", 00:28:21.894 "uuid": "66a5765f-21f4-5295-a031-7044d144ce73", 00:28:21.894 "is_configured": true, 00:28:21.894 "data_offset": 0, 00:28:21.894 "data_size": 65536 00:28:21.894 } 00:28:21.894 ] 00:28:21.894 }' 00:28:21.894 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:21.894 [2024-10-01 20:27:17.141439] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:28:22.152 [2024-10-01 20:27:17.149473] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:28:22.152 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:22.152 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:22.152 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:22.152 20:27:17 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:28:22.410 [2024-10-01 20:27:17.484090] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:28:22.668 [2024-10-01 20:27:17.697015] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:28:22.926 [2024-10-01 20:27:17.941809] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:28:22.926 [2024-10-01 20:27:17.943374] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:28:23.184 108.40 IOPS, 325.20 MiB/s [2024-10-01 20:27:18.184926] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:23.184 "name": "raid_bdev1", 00:28:23.184 "uuid": "a0828108-1a63-4e8f-9294-52cbf3f69ca9", 00:28:23.184 "strip_size_kb": 0, 00:28:23.184 "state": "online", 00:28:23.184 "raid_level": "raid1", 00:28:23.184 "superblock": false, 00:28:23.184 "num_base_bdevs": 4, 00:28:23.184 "num_base_bdevs_discovered": 3, 00:28:23.184 "num_base_bdevs_operational": 3, 00:28:23.184 "process": { 00:28:23.184 "type": "rebuild", 00:28:23.184 "target": "spare", 00:28:23.184 "progress": { 00:28:23.184 "blocks": 28672, 00:28:23.184 "percent": 43 00:28:23.184 } 00:28:23.184 }, 00:28:23.184 "base_bdevs_list": [ 00:28:23.184 { 00:28:23.184 "name": "spare", 00:28:23.184 "uuid": "af5e50a2-7b9f-5a01-ac8d-3a11fb14aab0", 00:28:23.184 "is_configured": true, 00:28:23.184 "data_offset": 0, 00:28:23.184 "data_size": 65536 00:28:23.184 }, 00:28:23.184 { 00:28:23.184 "name": null, 00:28:23.184 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:23.184 "is_configured": false, 00:28:23.184 "data_offset": 0, 00:28:23.184 "data_size": 65536 00:28:23.184 }, 00:28:23.184 { 00:28:23.184 "name": "BaseBdev3", 00:28:23.184 "uuid": "b46da91b-a0df-5787-9186-77651356325b", 00:28:23.184 "is_configured": true, 00:28:23.184 "data_offset": 0, 00:28:23.184 "data_size": 65536 00:28:23.184 }, 00:28:23.184 { 00:28:23.184 "name": "BaseBdev4", 00:28:23.184 "uuid": "66a5765f-21f4-5295-a031-7044d144ce73", 00:28:23.184 "is_configured": true, 00:28:23.184 "data_offset": 0, 00:28:23.184 "data_size": 65536 00:28:23.184 } 00:28:23.184 ] 00:28:23.184 }' 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:23.184 20:27:18 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:28:23.444 [2024-10-01 20:27:18.546273] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:28:23.702 [2024-10-01 20:27:18.876138] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:28:24.217 98.50 IOPS, 295.50 MiB/s 20:27:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:28:24.217 20:27:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:24.217 20:27:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:24.217 20:27:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:24.217 20:27:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:24.217 20:27:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:24.217 20:27:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:24.217 20:27:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:24.217 20:27:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:24.217 20:27:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:24.217 20:27:19 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:24.217 20:27:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:24.217 "name": "raid_bdev1", 00:28:24.218 "uuid": "a0828108-1a63-4e8f-9294-52cbf3f69ca9", 00:28:24.218 "strip_size_kb": 0, 00:28:24.218 "state": "online", 00:28:24.218 "raid_level": "raid1", 00:28:24.218 "superblock": false, 00:28:24.218 "num_base_bdevs": 4, 00:28:24.218 "num_base_bdevs_discovered": 3, 00:28:24.218 "num_base_bdevs_operational": 3, 00:28:24.218 "process": { 00:28:24.218 "type": "rebuild", 00:28:24.218 "target": "spare", 00:28:24.218 "progress": { 00:28:24.218 "blocks": 47104, 00:28:24.218 "percent": 71 00:28:24.218 } 00:28:24.218 }, 00:28:24.218 "base_bdevs_list": [ 00:28:24.218 { 00:28:24.218 "name": "spare", 00:28:24.218 "uuid": "af5e50a2-7b9f-5a01-ac8d-3a11fb14aab0", 00:28:24.218 "is_configured": true, 00:28:24.218 "data_offset": 0, 00:28:24.218 "data_size": 65536 00:28:24.218 }, 00:28:24.218 { 00:28:24.218 "name": null, 00:28:24.218 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:24.218 "is_configured": false, 00:28:24.218 "data_offset": 0, 00:28:24.218 "data_size": 65536 00:28:24.218 }, 00:28:24.218 { 00:28:24.218 "name": "BaseBdev3", 00:28:24.218 "uuid": "b46da91b-a0df-5787-9186-77651356325b", 00:28:24.218 "is_configured": true, 00:28:24.218 "data_offset": 0, 00:28:24.218 "data_size": 65536 00:28:24.218 }, 00:28:24.218 { 00:28:24.218 "name": "BaseBdev4", 00:28:24.218 "uuid": "66a5765f-21f4-5295-a031-7044d144ce73", 00:28:24.218 "is_configured": true, 00:28:24.218 "data_offset": 0, 00:28:24.218 "data_size": 65536 00:28:24.218 } 00:28:24.218 ] 00:28:24.218 }' 00:28:24.218 20:27:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:24.474 20:27:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:24.474 20:27:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:24.474 20:27:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:24.474 20:27:19 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:28:24.731 [2024-10-01 20:27:19.925606] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:28:25.247 90.00 IOPS, 270.00 MiB/s [2024-10-01 20:27:20.469634] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:25.505 [2024-10-01 20:27:20.569666] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:28:25.505 [2024-10-01 20:27:20.572928] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:25.505 "name": "raid_bdev1", 00:28:25.505 "uuid": "a0828108-1a63-4e8f-9294-52cbf3f69ca9", 00:28:25.505 "strip_size_kb": 0, 00:28:25.505 "state": "online", 00:28:25.505 "raid_level": "raid1", 00:28:25.505 "superblock": false, 00:28:25.505 "num_base_bdevs": 4, 00:28:25.505 "num_base_bdevs_discovered": 3, 00:28:25.505 "num_base_bdevs_operational": 3, 00:28:25.505 "process": { 00:28:25.505 "type": "rebuild", 00:28:25.505 "target": "spare", 00:28:25.505 "progress": { 00:28:25.505 "blocks": 65536, 00:28:25.505 "percent": 100 00:28:25.505 } 00:28:25.505 }, 00:28:25.505 "base_bdevs_list": [ 00:28:25.505 { 00:28:25.505 "name": "spare", 00:28:25.505 "uuid": "af5e50a2-7b9f-5a01-ac8d-3a11fb14aab0", 00:28:25.505 "is_configured": true, 00:28:25.505 "data_offset": 0, 00:28:25.505 "data_size": 65536 00:28:25.505 }, 00:28:25.505 { 00:28:25.505 "name": null, 00:28:25.505 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:25.505 "is_configured": false, 00:28:25.505 "data_offset": 0, 00:28:25.505 "data_size": 65536 00:28:25.505 }, 00:28:25.505 { 00:28:25.505 "name": "BaseBdev3", 00:28:25.505 "uuid": "b46da91b-a0df-5787-9186-77651356325b", 00:28:25.505 "is_configured": true, 00:28:25.505 "data_offset": 0, 00:28:25.505 "data_size": 65536 00:28:25.505 }, 00:28:25.505 { 00:28:25.505 "name": "BaseBdev4", 00:28:25.505 "uuid": "66a5765f-21f4-5295-a031-7044d144ce73", 00:28:25.505 "is_configured": true, 00:28:25.505 "data_offset": 0, 00:28:25.505 "data_size": 65536 00:28:25.505 } 00:28:25.505 ] 00:28:25.505 }' 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:25.505 20:27:20 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:28:26.638 82.62 IOPS, 247.88 MiB/s 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:28:26.638 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:26.638 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:26.638 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:26.638 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:26.638 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:26.638 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:26.638 20:27:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:26.638 20:27:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:26.638 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:26.638 20:27:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:26.638 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:26.638 "name": "raid_bdev1", 00:28:26.638 "uuid": "a0828108-1a63-4e8f-9294-52cbf3f69ca9", 00:28:26.638 "strip_size_kb": 0, 00:28:26.638 "state": "online", 00:28:26.638 "raid_level": "raid1", 00:28:26.638 "superblock": false, 00:28:26.638 "num_base_bdevs": 4, 00:28:26.638 "num_base_bdevs_discovered": 3, 00:28:26.638 "num_base_bdevs_operational": 3, 00:28:26.638 "base_bdevs_list": [ 00:28:26.638 { 00:28:26.638 "name": "spare", 00:28:26.638 "uuid": "af5e50a2-7b9f-5a01-ac8d-3a11fb14aab0", 00:28:26.638 "is_configured": true, 00:28:26.638 "data_offset": 0, 00:28:26.638 "data_size": 65536 00:28:26.638 }, 00:28:26.638 { 00:28:26.638 "name": null, 00:28:26.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:26.638 "is_configured": false, 00:28:26.638 "data_offset": 0, 00:28:26.638 "data_size": 65536 00:28:26.638 }, 00:28:26.638 { 00:28:26.638 "name": "BaseBdev3", 00:28:26.638 "uuid": "b46da91b-a0df-5787-9186-77651356325b", 00:28:26.638 "is_configured": true, 00:28:26.638 "data_offset": 0, 00:28:26.638 "data_size": 65536 00:28:26.638 }, 00:28:26.638 { 00:28:26.638 "name": "BaseBdev4", 00:28:26.638 "uuid": "66a5765f-21f4-5295-a031-7044d144ce73", 00:28:26.638 "is_configured": true, 00:28:26.638 "data_offset": 0, 00:28:26.638 "data_size": 65536 00:28:26.638 } 00:28:26.638 ] 00:28:26.638 }' 00:28:26.638 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:26.638 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:28:26.638 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:26.897 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:28:26.897 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:28:26.897 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:26.897 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:26.897 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:28:26.897 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:28:26.897 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:26.897 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:26.897 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:26.897 20:27:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:26.897 20:27:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:26.897 20:27:21 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:26.897 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:26.897 "name": "raid_bdev1", 00:28:26.897 "uuid": "a0828108-1a63-4e8f-9294-52cbf3f69ca9", 00:28:26.897 "strip_size_kb": 0, 00:28:26.897 "state": "online", 00:28:26.897 "raid_level": "raid1", 00:28:26.897 "superblock": false, 00:28:26.897 "num_base_bdevs": 4, 00:28:26.897 "num_base_bdevs_discovered": 3, 00:28:26.897 "num_base_bdevs_operational": 3, 00:28:26.897 "base_bdevs_list": [ 00:28:26.897 { 00:28:26.897 "name": "spare", 00:28:26.897 "uuid": "af5e50a2-7b9f-5a01-ac8d-3a11fb14aab0", 00:28:26.897 "is_configured": true, 00:28:26.897 "data_offset": 0, 00:28:26.897 "data_size": 65536 00:28:26.897 }, 00:28:26.897 { 00:28:26.897 "name": null, 00:28:26.897 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:26.897 "is_configured": false, 00:28:26.897 "data_offset": 0, 00:28:26.897 "data_size": 65536 00:28:26.897 }, 00:28:26.897 { 00:28:26.897 "name": "BaseBdev3", 00:28:26.897 "uuid": "b46da91b-a0df-5787-9186-77651356325b", 00:28:26.897 "is_configured": true, 00:28:26.897 "data_offset": 0, 00:28:26.897 "data_size": 65536 00:28:26.897 }, 00:28:26.897 { 00:28:26.897 "name": "BaseBdev4", 00:28:26.897 "uuid": "66a5765f-21f4-5295-a031-7044d144ce73", 00:28:26.897 "is_configured": true, 00:28:26.897 "data_offset": 0, 00:28:26.897 "data_size": 65536 00:28:26.897 } 00:28:26.897 ] 00:28:26.897 }' 00:28:26.897 20:27:21 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:26.897 77.78 IOPS, 233.33 MiB/s 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:26.897 "name": "raid_bdev1", 00:28:26.897 "uuid": "a0828108-1a63-4e8f-9294-52cbf3f69ca9", 00:28:26.897 "strip_size_kb": 0, 00:28:26.897 "state": "online", 00:28:26.897 "raid_level": "raid1", 00:28:26.897 "superblock": false, 00:28:26.897 "num_base_bdevs": 4, 00:28:26.897 "num_base_bdevs_discovered": 3, 00:28:26.897 "num_base_bdevs_operational": 3, 00:28:26.897 "base_bdevs_list": [ 00:28:26.897 { 00:28:26.897 "name": "spare", 00:28:26.897 "uuid": "af5e50a2-7b9f-5a01-ac8d-3a11fb14aab0", 00:28:26.897 "is_configured": true, 00:28:26.897 "data_offset": 0, 00:28:26.897 "data_size": 65536 00:28:26.897 }, 00:28:26.897 { 00:28:26.897 "name": null, 00:28:26.897 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:26.897 "is_configured": false, 00:28:26.897 "data_offset": 0, 00:28:26.897 "data_size": 65536 00:28:26.897 }, 00:28:26.897 { 00:28:26.897 "name": "BaseBdev3", 00:28:26.897 "uuid": "b46da91b-a0df-5787-9186-77651356325b", 00:28:26.897 "is_configured": true, 00:28:26.897 "data_offset": 0, 00:28:26.897 "data_size": 65536 00:28:26.897 }, 00:28:26.897 { 00:28:26.897 "name": "BaseBdev4", 00:28:26.897 "uuid": "66a5765f-21f4-5295-a031-7044d144ce73", 00:28:26.897 "is_configured": true, 00:28:26.897 "data_offset": 0, 00:28:26.897 "data_size": 65536 00:28:26.897 } 00:28:26.897 ] 00:28:26.897 }' 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:26.897 20:27:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:27.465 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:28:27.465 20:27:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:27.465 20:27:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:27.465 [2024-10-01 20:27:22.608388] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:27.465 [2024-10-01 20:27:22.608571] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:27.465 00:28:27.465 Latency(us) 00:28:27.465 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:27.465 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:28:27.465 raid_bdev1 : 9.66 74.77 224.32 0.00 0.00 19027.46 294.17 113436.86 00:28:27.465 =================================================================================================================== 00:28:27.465 Total : 74.77 224.32 0.00 0.00 19027.46 294.17 113436.86 00:28:27.465 { 00:28:27.465 "results": [ 00:28:27.465 { 00:28:27.465 "job": "raid_bdev1", 00:28:27.465 "core_mask": "0x1", 00:28:27.465 "workload": "randrw", 00:28:27.465 "percentage": 50, 00:28:27.465 "status": "finished", 00:28:27.465 "queue_depth": 2, 00:28:27.465 "io_size": 3145728, 00:28:27.465 "runtime": 9.655707, 00:28:27.465 "iops": 74.77443132853969, 00:28:27.465 "mibps": 224.32329398561907, 00:28:27.465 "io_failed": 0, 00:28:27.465 "io_timeout": 0, 00:28:27.465 "avg_latency_us": 19027.462059934525, 00:28:27.465 "min_latency_us": 294.16727272727275, 00:28:27.465 "max_latency_us": 113436.85818181818 00:28:27.465 } 00:28:27.465 ], 00:28:27.465 "core_count": 1 00:28:27.465 } 00:28:27.465 [2024-10-01 20:27:22.696762] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:27.465 [2024-10-01 20:27:22.696848] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:27.465 [2024-10-01 20:27:22.696994] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:27.465 [2024-10-01 20:27:22.697020] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:28:27.465 20:27:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:27.465 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:27.465 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:28:27.465 20:27:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:27.465 20:27:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:27.465 20:27:22 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:27.723 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:28:27.723 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:28:27.723 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:28:27.723 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:28:27.723 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:28:27.723 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:28:27.723 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:28:27.723 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:28:27.723 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:28:27.723 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:28:27.723 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:28:27.723 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:28:27.723 20:27:22 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:28:27.982 /dev/nbd0 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:28:27.982 1+0 records in 00:28:27.982 1+0 records out 00:28:27.982 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000379868 s, 10.8 MB/s 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@728 -- # continue 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:28:27.982 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:28:28.240 /dev/nbd1 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:28:28.240 1+0 records in 00:28:28.240 1+0 records out 00:28:28.240 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00034386 s, 11.9 MB/s 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:28:28.240 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:28:28.499 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:28:28.499 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:28:28.499 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:28:28.499 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:28:28.499 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:28:28.499 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:28:28.499 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:28:28.758 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:28:28.758 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:28:28.758 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:28:28.758 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:28:28.759 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:28:28.759 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:28:28.759 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:28:28.759 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:28:28.759 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:28:28.759 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:28:28.759 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:28:28.759 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:28:28.759 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:28:28.759 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:28:28.759 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:28:28.759 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:28:28.759 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:28:28.759 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:28:28.759 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:28:28.759 20:27:23 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:28:29.018 /dev/nbd1 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:28:29.277 1+0 records in 00:28:29.277 1+0 records out 00:28:29.277 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000429305 s, 9.5 MB/s 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:28:29.277 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:28:29.278 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:28:29.278 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:28:29.278 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:28:29.278 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:28:29.278 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:28:29.278 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:28:29.278 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:28:29.278 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:28:29.536 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:28:29.536 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:28:29.536 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:28:29.536 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:28:29.536 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:28:29.536 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:28:29.537 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:28:29.537 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:28:29.537 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:28:29.537 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:28:29.537 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:28:29.537 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:28:29.537 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:28:29.537 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:28:29.537 20:27:24 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:28:29.796 20:27:25 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:28:29.796 20:27:25 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:28:29.796 20:27:25 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:28:29.796 20:27:25 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:28:29.796 20:27:25 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:28:29.796 20:27:25 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:28:29.796 20:27:25 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:28:29.796 20:27:25 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:28:29.796 20:27:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:28:29.796 20:27:25 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 79904 00:28:29.796 20:27:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@950 -- # '[' -z 79904 ']' 00:28:29.796 20:27:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # kill -0 79904 00:28:29.796 20:27:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@955 -- # uname 00:28:29.796 20:27:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:28:29.796 20:27:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 79904 00:28:30.055 killing process with pid 79904 00:28:30.055 Received shutdown signal, test time was about 12.036110 seconds 00:28:30.055 00:28:30.055 Latency(us) 00:28:30.055 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:30.055 =================================================================================================================== 00:28:30.055 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:28:30.055 20:27:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:28:30.055 20:27:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:28:30.055 20:27:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@968 -- # echo 'killing process with pid 79904' 00:28:30.055 20:27:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@969 -- # kill 79904 00:28:30.055 [2024-10-01 20:27:25.056725] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:28:30.056 20:27:25 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@974 -- # wait 79904 00:28:30.374 [2024-10-01 20:27:25.439331] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:28:32.277 00:28:32.277 real 0m16.446s 00:28:32.277 user 0m21.148s 00:28:32.277 sys 0m1.972s 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:28:32.277 ************************************ 00:28:32.277 END TEST raid_rebuild_test_io 00:28:32.277 ************************************ 00:28:32.277 20:27:27 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 4 true true true 00:28:32.277 20:27:27 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:28:32.277 20:27:27 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:28:32.277 20:27:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:28:32.277 ************************************ 00:28:32.277 START TEST raid_rebuild_test_sb_io 00:28:32.277 ************************************ 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 4 true true true 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=80349 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 80349 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@831 -- # '[' -z 80349 ']' 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:32.277 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@836 -- # local max_retries=100 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # xtrace_disable 00:28:32.277 20:27:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:32.278 [2024-10-01 20:27:27.426888] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:28:32.278 [2024-10-01 20:27:27.427061] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid80349 ] 00:28:32.278 I/O size of 3145728 is greater than zero copy threshold (65536). 00:28:32.278 Zero copy mechanism will not be used. 00:28:32.536 [2024-10-01 20:27:27.594630] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:32.795 [2024-10-01 20:27:27.859201] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:28:33.054 [2024-10-01 20:27:28.063327] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:33.054 [2024-10-01 20:27:28.063403] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:33.312 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:28:33.312 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # return 0 00:28:33.312 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:28:33.312 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:28:33.312 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:33.312 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:33.312 BaseBdev1_malloc 00:28:33.312 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:33.312 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:28:33.312 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:33.312 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:33.312 [2024-10-01 20:27:28.528936] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:28:33.312 [2024-10-01 20:27:28.529024] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:33.312 [2024-10-01 20:27:28.529063] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:28:33.312 [2024-10-01 20:27:28.529089] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:33.312 [2024-10-01 20:27:28.532134] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:33.312 [2024-10-01 20:27:28.532183] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:28:33.312 BaseBdev1 00:28:33.312 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:33.312 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:28:33.312 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:28:33.312 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:33.312 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:33.571 BaseBdev2_malloc 00:28:33.571 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:33.571 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:28:33.571 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:33.571 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:33.571 [2024-10-01 20:27:28.585419] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:28:33.571 [2024-10-01 20:27:28.585508] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:33.571 [2024-10-01 20:27:28.585551] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:28:33.571 [2024-10-01 20:27:28.585572] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:33.571 [2024-10-01 20:27:28.588589] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:33.571 [2024-10-01 20:27:28.588640] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:28:33.571 BaseBdev2 00:28:33.571 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:33.571 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:28:33.571 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:28:33.571 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:33.571 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:33.571 BaseBdev3_malloc 00:28:33.571 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:33.571 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:28:33.571 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:33.571 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:33.571 [2024-10-01 20:27:28.642175] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:28:33.571 [2024-10-01 20:27:28.642254] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:33.571 [2024-10-01 20:27:28.642293] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:28:33.571 [2024-10-01 20:27:28.642313] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:33.571 [2024-10-01 20:27:28.645307] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:33.571 [2024-10-01 20:27:28.645357] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:28:33.571 BaseBdev3 00:28:33.571 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:33.571 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:28:33.571 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:33.572 BaseBdev4_malloc 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:33.572 [2024-10-01 20:27:28.698584] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:28:33.572 [2024-10-01 20:27:28.698673] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:33.572 [2024-10-01 20:27:28.698708] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:28:33.572 [2024-10-01 20:27:28.698746] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:33.572 [2024-10-01 20:27:28.701795] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:33.572 [2024-10-01 20:27:28.701846] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:28:33.572 BaseBdev4 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:33.572 spare_malloc 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:33.572 spare_delay 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:33.572 [2024-10-01 20:27:28.762964] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:28:33.572 [2024-10-01 20:27:28.763044] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:33.572 [2024-10-01 20:27:28.763079] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:28:33.572 [2024-10-01 20:27:28.763099] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:33.572 [2024-10-01 20:27:28.766062] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:33.572 [2024-10-01 20:27:28.766113] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:28:33.572 spare 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:33.572 [2024-10-01 20:27:28.775076] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:33.572 [2024-10-01 20:27:28.777627] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:33.572 [2024-10-01 20:27:28.777764] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:33.572 [2024-10-01 20:27:28.777856] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:28:33.572 [2024-10-01 20:27:28.778134] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:28:33.572 [2024-10-01 20:27:28.778164] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:28:33.572 [2024-10-01 20:27:28.778534] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:28:33.572 [2024-10-01 20:27:28.778812] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:28:33.572 [2024-10-01 20:27:28.778836] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:28:33.572 [2024-10-01 20:27:28.779110] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:33.572 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:33.831 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:33.831 "name": "raid_bdev1", 00:28:33.831 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:33.831 "strip_size_kb": 0, 00:28:33.831 "state": "online", 00:28:33.831 "raid_level": "raid1", 00:28:33.831 "superblock": true, 00:28:33.831 "num_base_bdevs": 4, 00:28:33.831 "num_base_bdevs_discovered": 4, 00:28:33.831 "num_base_bdevs_operational": 4, 00:28:33.831 "base_bdevs_list": [ 00:28:33.831 { 00:28:33.831 "name": "BaseBdev1", 00:28:33.831 "uuid": "0f8d0f52-b249-5d4d-8af8-326e7eb5a4dc", 00:28:33.831 "is_configured": true, 00:28:33.831 "data_offset": 2048, 00:28:33.831 "data_size": 63488 00:28:33.831 }, 00:28:33.831 { 00:28:33.831 "name": "BaseBdev2", 00:28:33.831 "uuid": "668413ee-bfeb-5686-ae7a-3648f0afa11a", 00:28:33.831 "is_configured": true, 00:28:33.831 "data_offset": 2048, 00:28:33.831 "data_size": 63488 00:28:33.831 }, 00:28:33.831 { 00:28:33.831 "name": "BaseBdev3", 00:28:33.831 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:33.831 "is_configured": true, 00:28:33.831 "data_offset": 2048, 00:28:33.831 "data_size": 63488 00:28:33.831 }, 00:28:33.831 { 00:28:33.831 "name": "BaseBdev4", 00:28:33.831 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:33.831 "is_configured": true, 00:28:33.831 "data_offset": 2048, 00:28:33.831 "data_size": 63488 00:28:33.831 } 00:28:33.831 ] 00:28:33.831 }' 00:28:33.831 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:33.831 20:27:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:34.090 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:28:34.090 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:28:34.090 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:34.090 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:34.090 [2024-10-01 20:27:29.283604] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:28:34.090 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:34.090 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:28:34.090 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:34.090 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:34.090 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:34.090 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:28:34.090 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:34.349 [2024-10-01 20:27:29.379240] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:34.349 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:34.349 "name": "raid_bdev1", 00:28:34.349 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:34.349 "strip_size_kb": 0, 00:28:34.349 "state": "online", 00:28:34.349 "raid_level": "raid1", 00:28:34.349 "superblock": true, 00:28:34.349 "num_base_bdevs": 4, 00:28:34.349 "num_base_bdevs_discovered": 3, 00:28:34.349 "num_base_bdevs_operational": 3, 00:28:34.349 "base_bdevs_list": [ 00:28:34.349 { 00:28:34.349 "name": null, 00:28:34.349 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:34.349 "is_configured": false, 00:28:34.349 "data_offset": 0, 00:28:34.349 "data_size": 63488 00:28:34.349 }, 00:28:34.349 { 00:28:34.349 "name": "BaseBdev2", 00:28:34.349 "uuid": "668413ee-bfeb-5686-ae7a-3648f0afa11a", 00:28:34.349 "is_configured": true, 00:28:34.349 "data_offset": 2048, 00:28:34.349 "data_size": 63488 00:28:34.349 }, 00:28:34.350 { 00:28:34.350 "name": "BaseBdev3", 00:28:34.350 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:34.350 "is_configured": true, 00:28:34.350 "data_offset": 2048, 00:28:34.350 "data_size": 63488 00:28:34.350 }, 00:28:34.350 { 00:28:34.350 "name": "BaseBdev4", 00:28:34.350 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:34.350 "is_configured": true, 00:28:34.350 "data_offset": 2048, 00:28:34.350 "data_size": 63488 00:28:34.350 } 00:28:34.350 ] 00:28:34.350 }' 00:28:34.350 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:34.350 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:34.350 [2024-10-01 20:27:29.491890] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:28:34.350 I/O size of 3145728 is greater than zero copy threshold (65536). 00:28:34.350 Zero copy mechanism will not be used. 00:28:34.350 Running I/O for 60 seconds... 00:28:34.917 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:28:34.917 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:34.917 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:34.917 [2024-10-01 20:27:29.915416] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:34.917 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:34.917 20:27:29 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:28:34.917 [2024-10-01 20:27:29.995269] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000062f0 00:28:34.917 [2024-10-01 20:27:29.998223] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:28:34.917 [2024-10-01 20:27:30.130460] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:28:34.917 [2024-10-01 20:27:30.132142] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:28:35.177 [2024-10-01 20:27:30.347612] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:28:35.177 [2024-10-01 20:27:30.348673] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:28:35.437 133.00 IOPS, 399.00 MiB/s [2024-10-01 20:27:30.687508] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:28:36.005 20:27:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:36.005 20:27:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:36.005 20:27:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:36.005 20:27:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:36.005 20:27:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:36.005 20:27:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:36.005 20:27:30 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:36.006 20:27:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:36.006 20:27:30 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:36.006 "name": "raid_bdev1", 00:28:36.006 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:36.006 "strip_size_kb": 0, 00:28:36.006 "state": "online", 00:28:36.006 "raid_level": "raid1", 00:28:36.006 "superblock": true, 00:28:36.006 "num_base_bdevs": 4, 00:28:36.006 "num_base_bdevs_discovered": 4, 00:28:36.006 "num_base_bdevs_operational": 4, 00:28:36.006 "process": { 00:28:36.006 "type": "rebuild", 00:28:36.006 "target": "spare", 00:28:36.006 "progress": { 00:28:36.006 "blocks": 10240, 00:28:36.006 "percent": 16 00:28:36.006 } 00:28:36.006 }, 00:28:36.006 "base_bdevs_list": [ 00:28:36.006 { 00:28:36.006 "name": "spare", 00:28:36.006 "uuid": "a4791e79-ff6a-5a2c-88ad-00ae2225591b", 00:28:36.006 "is_configured": true, 00:28:36.006 "data_offset": 2048, 00:28:36.006 "data_size": 63488 00:28:36.006 }, 00:28:36.006 { 00:28:36.006 "name": "BaseBdev2", 00:28:36.006 "uuid": "668413ee-bfeb-5686-ae7a-3648f0afa11a", 00:28:36.006 "is_configured": true, 00:28:36.006 "data_offset": 2048, 00:28:36.006 "data_size": 63488 00:28:36.006 }, 00:28:36.006 { 00:28:36.006 "name": "BaseBdev3", 00:28:36.006 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:36.006 "is_configured": true, 00:28:36.006 "data_offset": 2048, 00:28:36.006 "data_size": 63488 00:28:36.006 }, 00:28:36.006 { 00:28:36.006 "name": "BaseBdev4", 00:28:36.006 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:36.006 "is_configured": true, 00:28:36.006 "data_offset": 2048, 00:28:36.006 "data_size": 63488 00:28:36.006 } 00:28:36.006 ] 00:28:36.006 }' 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:36.006 [2024-10-01 20:27:31.169832] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:36.006 [2024-10-01 20:27:31.194873] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:28:36.006 [2024-10-01 20:27:31.206510] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:36.006 [2024-10-01 20:27:31.206626] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:36.006 [2024-10-01 20:27:31.206649] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:28:36.006 [2024-10-01 20:27:31.240055] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000006220 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:36.006 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:36.309 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:36.309 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:36.309 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:36.309 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:36.309 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:36.309 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:36.309 "name": "raid_bdev1", 00:28:36.309 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:36.309 "strip_size_kb": 0, 00:28:36.309 "state": "online", 00:28:36.309 "raid_level": "raid1", 00:28:36.309 "superblock": true, 00:28:36.309 "num_base_bdevs": 4, 00:28:36.309 "num_base_bdevs_discovered": 3, 00:28:36.309 "num_base_bdevs_operational": 3, 00:28:36.309 "base_bdevs_list": [ 00:28:36.309 { 00:28:36.309 "name": null, 00:28:36.309 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:36.309 "is_configured": false, 00:28:36.309 "data_offset": 0, 00:28:36.309 "data_size": 63488 00:28:36.309 }, 00:28:36.309 { 00:28:36.309 "name": "BaseBdev2", 00:28:36.309 "uuid": "668413ee-bfeb-5686-ae7a-3648f0afa11a", 00:28:36.309 "is_configured": true, 00:28:36.309 "data_offset": 2048, 00:28:36.309 "data_size": 63488 00:28:36.309 }, 00:28:36.309 { 00:28:36.309 "name": "BaseBdev3", 00:28:36.309 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:36.309 "is_configured": true, 00:28:36.309 "data_offset": 2048, 00:28:36.309 "data_size": 63488 00:28:36.309 }, 00:28:36.309 { 00:28:36.309 "name": "BaseBdev4", 00:28:36.309 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:36.309 "is_configured": true, 00:28:36.309 "data_offset": 2048, 00:28:36.309 "data_size": 63488 00:28:36.309 } 00:28:36.309 ] 00:28:36.309 }' 00:28:36.309 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:36.309 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:36.566 126.50 IOPS, 379.50 MiB/s 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:36.566 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:36.566 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:28:36.566 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:28:36.566 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:36.566 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:36.566 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:36.566 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:36.566 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:36.824 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:36.824 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:36.824 "name": "raid_bdev1", 00:28:36.824 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:36.824 "strip_size_kb": 0, 00:28:36.824 "state": "online", 00:28:36.824 "raid_level": "raid1", 00:28:36.824 "superblock": true, 00:28:36.824 "num_base_bdevs": 4, 00:28:36.824 "num_base_bdevs_discovered": 3, 00:28:36.824 "num_base_bdevs_operational": 3, 00:28:36.824 "base_bdevs_list": [ 00:28:36.824 { 00:28:36.824 "name": null, 00:28:36.824 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:36.824 "is_configured": false, 00:28:36.824 "data_offset": 0, 00:28:36.825 "data_size": 63488 00:28:36.825 }, 00:28:36.825 { 00:28:36.825 "name": "BaseBdev2", 00:28:36.825 "uuid": "668413ee-bfeb-5686-ae7a-3648f0afa11a", 00:28:36.825 "is_configured": true, 00:28:36.825 "data_offset": 2048, 00:28:36.825 "data_size": 63488 00:28:36.825 }, 00:28:36.825 { 00:28:36.825 "name": "BaseBdev3", 00:28:36.825 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:36.825 "is_configured": true, 00:28:36.825 "data_offset": 2048, 00:28:36.825 "data_size": 63488 00:28:36.825 }, 00:28:36.825 { 00:28:36.825 "name": "BaseBdev4", 00:28:36.825 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:36.825 "is_configured": true, 00:28:36.825 "data_offset": 2048, 00:28:36.825 "data_size": 63488 00:28:36.825 } 00:28:36.825 ] 00:28:36.825 }' 00:28:36.825 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:36.825 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:28:36.825 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:36.825 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:28:36.825 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:28:36.825 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:36.825 20:27:31 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:36.825 [2024-10-01 20:27:31.967760] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:36.825 20:27:32 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:36.825 20:27:32 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:28:36.825 [2024-10-01 20:27:32.075708] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:28:37.084 [2024-10-01 20:27:32.078535] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:28:37.084 [2024-10-01 20:27:32.205976] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:28:37.343 [2024-10-01 20:27:32.461242] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:28:37.343 [2024-10-01 20:27:32.462317] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:28:37.602 139.33 IOPS, 418.00 MiB/s [2024-10-01 20:27:32.803514] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:28:37.861 [2024-10-01 20:27:32.927181] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:28:37.861 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:37.861 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:37.861 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:37.861 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:37.861 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:37.861 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:37.861 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:37.861 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:37.861 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:37.861 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:37.861 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:37.861 "name": "raid_bdev1", 00:28:37.861 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:37.861 "strip_size_kb": 0, 00:28:37.861 "state": "online", 00:28:37.861 "raid_level": "raid1", 00:28:37.861 "superblock": true, 00:28:37.861 "num_base_bdevs": 4, 00:28:37.861 "num_base_bdevs_discovered": 4, 00:28:37.861 "num_base_bdevs_operational": 4, 00:28:37.861 "process": { 00:28:37.861 "type": "rebuild", 00:28:37.861 "target": "spare", 00:28:37.861 "progress": { 00:28:37.861 "blocks": 12288, 00:28:37.861 "percent": 19 00:28:37.861 } 00:28:37.861 }, 00:28:37.861 "base_bdevs_list": [ 00:28:37.861 { 00:28:37.861 "name": "spare", 00:28:37.861 "uuid": "a4791e79-ff6a-5a2c-88ad-00ae2225591b", 00:28:37.861 "is_configured": true, 00:28:37.861 "data_offset": 2048, 00:28:37.861 "data_size": 63488 00:28:37.861 }, 00:28:37.861 { 00:28:37.861 "name": "BaseBdev2", 00:28:37.861 "uuid": "668413ee-bfeb-5686-ae7a-3648f0afa11a", 00:28:37.861 "is_configured": true, 00:28:37.861 "data_offset": 2048, 00:28:37.861 "data_size": 63488 00:28:37.861 }, 00:28:37.861 { 00:28:37.861 "name": "BaseBdev3", 00:28:37.861 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:37.861 "is_configured": true, 00:28:37.861 "data_offset": 2048, 00:28:37.861 "data_size": 63488 00:28:37.861 }, 00:28:37.861 { 00:28:37.861 "name": "BaseBdev4", 00:28:37.861 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:37.861 "is_configured": true, 00:28:37.861 "data_offset": 2048, 00:28:37.861 "data_size": 63488 00:28:37.861 } 00:28:37.861 ] 00:28:37.861 }' 00:28:37.861 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:28:38.120 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:38.120 [2024-10-01 20:27:33.214511] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:28:38.120 [2024-10-01 20:27:33.236284] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:28:38.120 [2024-10-01 20:27:33.236688] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:28:38.120 [2024-10-01 20:27:33.346401] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000006220 00:28:38.120 [2024-10-01 20:27:33.346500] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d0000063c0 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:38.120 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:38.379 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:38.379 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:38.379 "name": "raid_bdev1", 00:28:38.379 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:38.379 "strip_size_kb": 0, 00:28:38.379 "state": "online", 00:28:38.379 "raid_level": "raid1", 00:28:38.379 "superblock": true, 00:28:38.379 "num_base_bdevs": 4, 00:28:38.379 "num_base_bdevs_discovered": 3, 00:28:38.379 "num_base_bdevs_operational": 3, 00:28:38.379 "process": { 00:28:38.379 "type": "rebuild", 00:28:38.379 "target": "spare", 00:28:38.379 "progress": { 00:28:38.379 "blocks": 16384, 00:28:38.379 "percent": 25 00:28:38.379 } 00:28:38.379 }, 00:28:38.379 "base_bdevs_list": [ 00:28:38.379 { 00:28:38.379 "name": "spare", 00:28:38.379 "uuid": "a4791e79-ff6a-5a2c-88ad-00ae2225591b", 00:28:38.379 "is_configured": true, 00:28:38.379 "data_offset": 2048, 00:28:38.379 "data_size": 63488 00:28:38.379 }, 00:28:38.379 { 00:28:38.379 "name": null, 00:28:38.379 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:38.379 "is_configured": false, 00:28:38.379 "data_offset": 0, 00:28:38.379 "data_size": 63488 00:28:38.379 }, 00:28:38.379 { 00:28:38.379 "name": "BaseBdev3", 00:28:38.379 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:38.379 "is_configured": true, 00:28:38.379 "data_offset": 2048, 00:28:38.379 "data_size": 63488 00:28:38.379 }, 00:28:38.379 { 00:28:38.379 "name": "BaseBdev4", 00:28:38.379 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:38.379 "is_configured": true, 00:28:38.379 "data_offset": 2048, 00:28:38.379 "data_size": 63488 00:28:38.379 } 00:28:38.379 ] 00:28:38.379 }' 00:28:38.379 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:38.380 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:38.380 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:38.380 131.00 IOPS, 393.00 MiB/s 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:38.380 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=592 00:28:38.380 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:28:38.380 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:38.380 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:38.380 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:38.380 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:38.380 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:38.380 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:38.380 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:38.380 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:38.380 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:38.380 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:38.380 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:38.380 "name": "raid_bdev1", 00:28:38.380 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:38.380 "strip_size_kb": 0, 00:28:38.380 "state": "online", 00:28:38.380 "raid_level": "raid1", 00:28:38.380 "superblock": true, 00:28:38.380 "num_base_bdevs": 4, 00:28:38.380 "num_base_bdevs_discovered": 3, 00:28:38.380 "num_base_bdevs_operational": 3, 00:28:38.380 "process": { 00:28:38.380 "type": "rebuild", 00:28:38.380 "target": "spare", 00:28:38.380 "progress": { 00:28:38.380 "blocks": 18432, 00:28:38.380 "percent": 29 00:28:38.380 } 00:28:38.380 }, 00:28:38.380 "base_bdevs_list": [ 00:28:38.380 { 00:28:38.380 "name": "spare", 00:28:38.380 "uuid": "a4791e79-ff6a-5a2c-88ad-00ae2225591b", 00:28:38.380 "is_configured": true, 00:28:38.380 "data_offset": 2048, 00:28:38.380 "data_size": 63488 00:28:38.380 }, 00:28:38.380 { 00:28:38.380 "name": null, 00:28:38.380 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:38.380 "is_configured": false, 00:28:38.380 "data_offset": 0, 00:28:38.380 "data_size": 63488 00:28:38.380 }, 00:28:38.380 { 00:28:38.380 "name": "BaseBdev3", 00:28:38.380 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:38.380 "is_configured": true, 00:28:38.380 "data_offset": 2048, 00:28:38.380 "data_size": 63488 00:28:38.380 }, 00:28:38.380 { 00:28:38.380 "name": "BaseBdev4", 00:28:38.380 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:38.380 "is_configured": true, 00:28:38.380 "data_offset": 2048, 00:28:38.380 "data_size": 63488 00:28:38.380 } 00:28:38.380 ] 00:28:38.380 }' 00:28:38.380 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:38.639 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:38.639 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:38.639 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:38.639 20:27:33 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:28:38.898 [2024-10-01 20:27:33.922688] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:28:38.898 [2024-10-01 20:27:33.923874] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:28:39.157 [2024-10-01 20:27:34.386511] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:28:39.415 114.40 IOPS, 343.20 MiB/s [2024-10-01 20:27:34.598049] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:28:39.674 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:28:39.674 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:39.674 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:39.674 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:39.674 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:39.675 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:39.675 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:39.675 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:39.675 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:39.675 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:39.675 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:39.675 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:39.675 "name": "raid_bdev1", 00:28:39.675 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:39.675 "strip_size_kb": 0, 00:28:39.675 "state": "online", 00:28:39.675 "raid_level": "raid1", 00:28:39.675 "superblock": true, 00:28:39.675 "num_base_bdevs": 4, 00:28:39.675 "num_base_bdevs_discovered": 3, 00:28:39.675 "num_base_bdevs_operational": 3, 00:28:39.675 "process": { 00:28:39.675 "type": "rebuild", 00:28:39.675 "target": "spare", 00:28:39.675 "progress": { 00:28:39.675 "blocks": 34816, 00:28:39.675 "percent": 54 00:28:39.675 } 00:28:39.675 }, 00:28:39.675 "base_bdevs_list": [ 00:28:39.675 { 00:28:39.675 "name": "spare", 00:28:39.675 "uuid": "a4791e79-ff6a-5a2c-88ad-00ae2225591b", 00:28:39.675 "is_configured": true, 00:28:39.675 "data_offset": 2048, 00:28:39.675 "data_size": 63488 00:28:39.675 }, 00:28:39.675 { 00:28:39.675 "name": null, 00:28:39.675 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:39.675 "is_configured": false, 00:28:39.675 "data_offset": 0, 00:28:39.675 "data_size": 63488 00:28:39.675 }, 00:28:39.675 { 00:28:39.675 "name": "BaseBdev3", 00:28:39.675 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:39.675 "is_configured": true, 00:28:39.675 "data_offset": 2048, 00:28:39.675 "data_size": 63488 00:28:39.675 }, 00:28:39.675 { 00:28:39.675 "name": "BaseBdev4", 00:28:39.675 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:39.675 "is_configured": true, 00:28:39.675 "data_offset": 2048, 00:28:39.675 "data_size": 63488 00:28:39.675 } 00:28:39.675 ] 00:28:39.675 }' 00:28:39.675 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:39.675 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:39.675 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:39.675 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:39.675 20:27:34 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:28:40.243 [2024-10-01 20:27:35.440229] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:28:40.780 103.67 IOPS, 311.00 MiB/s 20:27:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:28:40.780 20:27:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:40.780 20:27:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:40.780 20:27:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:40.780 20:27:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:40.780 20:27:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:40.780 20:27:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:40.780 20:27:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:40.780 20:27:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:40.780 20:27:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:40.780 20:27:35 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:40.780 20:27:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:40.780 "name": "raid_bdev1", 00:28:40.780 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:40.780 "strip_size_kb": 0, 00:28:40.780 "state": "online", 00:28:40.780 "raid_level": "raid1", 00:28:40.780 "superblock": true, 00:28:40.780 "num_base_bdevs": 4, 00:28:40.780 "num_base_bdevs_discovered": 3, 00:28:40.780 "num_base_bdevs_operational": 3, 00:28:40.780 "process": { 00:28:40.780 "type": "rebuild", 00:28:40.780 "target": "spare", 00:28:40.780 "progress": { 00:28:40.780 "blocks": 53248, 00:28:40.780 "percent": 83 00:28:40.780 } 00:28:40.780 }, 00:28:40.780 "base_bdevs_list": [ 00:28:40.780 { 00:28:40.780 "name": "spare", 00:28:40.780 "uuid": "a4791e79-ff6a-5a2c-88ad-00ae2225591b", 00:28:40.780 "is_configured": true, 00:28:40.780 "data_offset": 2048, 00:28:40.780 "data_size": 63488 00:28:40.780 }, 00:28:40.780 { 00:28:40.780 "name": null, 00:28:40.780 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:40.780 "is_configured": false, 00:28:40.780 "data_offset": 0, 00:28:40.780 "data_size": 63488 00:28:40.780 }, 00:28:40.780 { 00:28:40.780 "name": "BaseBdev3", 00:28:40.780 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:40.780 "is_configured": true, 00:28:40.780 "data_offset": 2048, 00:28:40.780 "data_size": 63488 00:28:40.780 }, 00:28:40.780 { 00:28:40.780 "name": "BaseBdev4", 00:28:40.780 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:40.780 "is_configured": true, 00:28:40.780 "data_offset": 2048, 00:28:40.780 "data_size": 63488 00:28:40.780 } 00:28:40.780 ] 00:28:40.780 }' 00:28:40.780 20:27:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:40.780 20:27:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:40.780 20:27:35 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:40.780 20:27:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:40.780 20:27:36 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:28:41.346 [2024-10-01 20:27:36.359234] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:28:41.346 [2024-10-01 20:27:36.459188] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:28:41.346 [2024-10-01 20:27:36.462648] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:41.913 93.29 IOPS, 279.86 MiB/s 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:28:41.913 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:41.913 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:41.913 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:41.913 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:41.913 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:41.913 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:41.913 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:41.913 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:41.913 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:41.913 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:41.913 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:41.913 "name": "raid_bdev1", 00:28:41.913 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:41.913 "strip_size_kb": 0, 00:28:41.913 "state": "online", 00:28:41.913 "raid_level": "raid1", 00:28:41.913 "superblock": true, 00:28:41.913 "num_base_bdevs": 4, 00:28:41.913 "num_base_bdevs_discovered": 3, 00:28:41.913 "num_base_bdevs_operational": 3, 00:28:41.913 "base_bdevs_list": [ 00:28:41.913 { 00:28:41.913 "name": "spare", 00:28:41.913 "uuid": "a4791e79-ff6a-5a2c-88ad-00ae2225591b", 00:28:41.913 "is_configured": true, 00:28:41.913 "data_offset": 2048, 00:28:41.913 "data_size": 63488 00:28:41.913 }, 00:28:41.913 { 00:28:41.913 "name": null, 00:28:41.913 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:41.913 "is_configured": false, 00:28:41.913 "data_offset": 0, 00:28:41.913 "data_size": 63488 00:28:41.913 }, 00:28:41.913 { 00:28:41.913 "name": "BaseBdev3", 00:28:41.913 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:41.913 "is_configured": true, 00:28:41.913 "data_offset": 2048, 00:28:41.913 "data_size": 63488 00:28:41.913 }, 00:28:41.913 { 00:28:41.913 "name": "BaseBdev4", 00:28:41.913 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:41.913 "is_configured": true, 00:28:41.913 "data_offset": 2048, 00:28:41.913 "data_size": 63488 00:28:41.913 } 00:28:41.913 ] 00:28:41.913 }' 00:28:41.913 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:41.913 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:28:41.913 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:42.173 "name": "raid_bdev1", 00:28:42.173 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:42.173 "strip_size_kb": 0, 00:28:42.173 "state": "online", 00:28:42.173 "raid_level": "raid1", 00:28:42.173 "superblock": true, 00:28:42.173 "num_base_bdevs": 4, 00:28:42.173 "num_base_bdevs_discovered": 3, 00:28:42.173 "num_base_bdevs_operational": 3, 00:28:42.173 "base_bdevs_list": [ 00:28:42.173 { 00:28:42.173 "name": "spare", 00:28:42.173 "uuid": "a4791e79-ff6a-5a2c-88ad-00ae2225591b", 00:28:42.173 "is_configured": true, 00:28:42.173 "data_offset": 2048, 00:28:42.173 "data_size": 63488 00:28:42.173 }, 00:28:42.173 { 00:28:42.173 "name": null, 00:28:42.173 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:42.173 "is_configured": false, 00:28:42.173 "data_offset": 0, 00:28:42.173 "data_size": 63488 00:28:42.173 }, 00:28:42.173 { 00:28:42.173 "name": "BaseBdev3", 00:28:42.173 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:42.173 "is_configured": true, 00:28:42.173 "data_offset": 2048, 00:28:42.173 "data_size": 63488 00:28:42.173 }, 00:28:42.173 { 00:28:42.173 "name": "BaseBdev4", 00:28:42.173 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:42.173 "is_configured": true, 00:28:42.173 "data_offset": 2048, 00:28:42.173 "data_size": 63488 00:28:42.173 } 00:28:42.173 ] 00:28:42.173 }' 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:42.173 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:42.173 "name": "raid_bdev1", 00:28:42.173 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:42.173 "strip_size_kb": 0, 00:28:42.173 "state": "online", 00:28:42.173 "raid_level": "raid1", 00:28:42.173 "superblock": true, 00:28:42.173 "num_base_bdevs": 4, 00:28:42.173 "num_base_bdevs_discovered": 3, 00:28:42.173 "num_base_bdevs_operational": 3, 00:28:42.173 "base_bdevs_list": [ 00:28:42.173 { 00:28:42.173 "name": "spare", 00:28:42.173 "uuid": "a4791e79-ff6a-5a2c-88ad-00ae2225591b", 00:28:42.173 "is_configured": true, 00:28:42.173 "data_offset": 2048, 00:28:42.173 "data_size": 63488 00:28:42.173 }, 00:28:42.173 { 00:28:42.173 "name": null, 00:28:42.173 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:42.173 "is_configured": false, 00:28:42.173 "data_offset": 0, 00:28:42.173 "data_size": 63488 00:28:42.174 }, 00:28:42.174 { 00:28:42.174 "name": "BaseBdev3", 00:28:42.174 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:42.174 "is_configured": true, 00:28:42.174 "data_offset": 2048, 00:28:42.174 "data_size": 63488 00:28:42.174 }, 00:28:42.174 { 00:28:42.174 "name": "BaseBdev4", 00:28:42.174 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:42.174 "is_configured": true, 00:28:42.174 "data_offset": 2048, 00:28:42.174 "data_size": 63488 00:28:42.174 } 00:28:42.174 ] 00:28:42.174 }' 00:28:42.174 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:42.174 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:42.692 87.00 IOPS, 261.00 MiB/s 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:28:42.692 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:42.692 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:42.692 [2024-10-01 20:27:37.823527] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:28:42.692 [2024-10-01 20:27:37.823562] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:28:42.692 00:28:42.692 Latency(us) 00:28:42.692 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:42.692 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:28:42.692 raid_bdev1 : 8.41 84.51 253.54 0.00 0.00 17258.22 292.31 122969.37 00:28:42.692 =================================================================================================================== 00:28:42.692 Total : 84.51 253.54 0.00 0.00 17258.22 292.31 122969.37 00:28:42.692 [2024-10-01 20:27:37.925898] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:42.692 { 00:28:42.692 "results": [ 00:28:42.692 { 00:28:42.692 "job": "raid_bdev1", 00:28:42.692 "core_mask": "0x1", 00:28:42.692 "workload": "randrw", 00:28:42.692 "percentage": 50, 00:28:42.692 "status": "finished", 00:28:42.692 "queue_depth": 2, 00:28:42.692 "io_size": 3145728, 00:28:42.692 "runtime": 8.412728, 00:28:42.692 "iops": 84.51479710267586, 00:28:42.692 "mibps": 253.54439130802757, 00:28:42.692 "io_failed": 0, 00:28:42.692 "io_timeout": 0, 00:28:42.692 "avg_latency_us": 17258.21659634318, 00:28:42.692 "min_latency_us": 292.30545454545455, 00:28:42.692 "max_latency_us": 122969.36727272728 00:28:42.692 } 00:28:42.692 ], 00:28:42.692 "core_count": 1 00:28:42.692 } 00:28:42.692 [2024-10-01 20:27:37.926109] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:42.692 [2024-10-01 20:27:37.926276] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:42.692 [2024-10-01 20:27:37.926297] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:28:42.692 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:42.692 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:28:42.692 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:42.692 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:42.692 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:42.950 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:42.950 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:28:42.951 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:28:42.951 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:28:42.951 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:28:42.951 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:28:42.951 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:28:42.951 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:28:42.951 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:28:42.951 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:28:42.951 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:28:42.951 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:28:42.951 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:28:42.951 20:27:37 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:28:43.210 /dev/nbd0 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:28:43.210 1+0 records in 00:28:43.210 1+0 records out 00:28:43.210 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000611736 s, 6.7 MB/s 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@728 -- # continue 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:28:43.210 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:28:43.469 /dev/nbd1 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:28:43.469 1+0 records in 00:28:43.469 1+0 records out 00:28:43.469 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000580224 s, 7.1 MB/s 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:28:43.469 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:28:43.728 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:28:43.728 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:28:43.728 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:28:43.728 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:28:43.728 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:28:43.728 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:28:43.728 20:27:38 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:28:43.988 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:28:44.248 /dev/nbd1 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:28:44.248 1+0 records in 00:28:44.248 1+0 records out 00:28:44.248 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000582437 s, 7.0 MB/s 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:28:44.248 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:28:44.507 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:28:44.507 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:28:44.507 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:28:44.507 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:28:44.507 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:28:44.507 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:28:44.507 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:28:44.507 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:28:44.507 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:28:44.507 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:28:44.507 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:28:44.507 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:28:44.507 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:28:44.507 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:28:44.507 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:28:44.768 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:28:44.768 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:28:44.768 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:28:44.768 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:28:44.768 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:28:44.768 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:28:44.768 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:28:44.768 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:28:44.768 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:28:44.769 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:28:44.769 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:44.769 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:44.769 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:44.769 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:28:44.769 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:44.769 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:44.769 [2024-10-01 20:27:39.983928] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:28:44.769 [2024-10-01 20:27:39.984157] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:44.769 [2024-10-01 20:27:39.984205] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:28:44.769 [2024-10-01 20:27:39.984221] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:44.769 [2024-10-01 20:27:39.987403] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:44.769 [2024-10-01 20:27:39.987576] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:28:44.769 [2024-10-01 20:27:39.987720] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:28:44.769 [2024-10-01 20:27:39.987837] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:44.769 [2024-10-01 20:27:39.988044] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:44.769 spare 00:28:44.769 [2024-10-01 20:27:39.988307] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:28:44.769 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:44.769 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:28:44.769 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:44.769 20:27:39 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:45.028 [2024-10-01 20:27:40.088452] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:28:45.028 [2024-10-01 20:27:40.088715] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:28:45.028 [2024-10-01 20:27:40.089351] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000037160 00:28:45.028 [2024-10-01 20:27:40.089628] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:28:45.028 [2024-10-01 20:27:40.089653] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:28:45.028 [2024-10-01 20:27:40.089912] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:45.028 "name": "raid_bdev1", 00:28:45.028 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:45.028 "strip_size_kb": 0, 00:28:45.028 "state": "online", 00:28:45.028 "raid_level": "raid1", 00:28:45.028 "superblock": true, 00:28:45.028 "num_base_bdevs": 4, 00:28:45.028 "num_base_bdevs_discovered": 3, 00:28:45.028 "num_base_bdevs_operational": 3, 00:28:45.028 "base_bdevs_list": [ 00:28:45.028 { 00:28:45.028 "name": "spare", 00:28:45.028 "uuid": "a4791e79-ff6a-5a2c-88ad-00ae2225591b", 00:28:45.028 "is_configured": true, 00:28:45.028 "data_offset": 2048, 00:28:45.028 "data_size": 63488 00:28:45.028 }, 00:28:45.028 { 00:28:45.028 "name": null, 00:28:45.028 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:45.028 "is_configured": false, 00:28:45.028 "data_offset": 2048, 00:28:45.028 "data_size": 63488 00:28:45.028 }, 00:28:45.028 { 00:28:45.028 "name": "BaseBdev3", 00:28:45.028 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:45.028 "is_configured": true, 00:28:45.028 "data_offset": 2048, 00:28:45.028 "data_size": 63488 00:28:45.028 }, 00:28:45.028 { 00:28:45.028 "name": "BaseBdev4", 00:28:45.028 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:45.028 "is_configured": true, 00:28:45.028 "data_offset": 2048, 00:28:45.028 "data_size": 63488 00:28:45.028 } 00:28:45.028 ] 00:28:45.028 }' 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:45.028 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:45.605 "name": "raid_bdev1", 00:28:45.605 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:45.605 "strip_size_kb": 0, 00:28:45.605 "state": "online", 00:28:45.605 "raid_level": "raid1", 00:28:45.605 "superblock": true, 00:28:45.605 "num_base_bdevs": 4, 00:28:45.605 "num_base_bdevs_discovered": 3, 00:28:45.605 "num_base_bdevs_operational": 3, 00:28:45.605 "base_bdevs_list": [ 00:28:45.605 { 00:28:45.605 "name": "spare", 00:28:45.605 "uuid": "a4791e79-ff6a-5a2c-88ad-00ae2225591b", 00:28:45.605 "is_configured": true, 00:28:45.605 "data_offset": 2048, 00:28:45.605 "data_size": 63488 00:28:45.605 }, 00:28:45.605 { 00:28:45.605 "name": null, 00:28:45.605 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:45.605 "is_configured": false, 00:28:45.605 "data_offset": 2048, 00:28:45.605 "data_size": 63488 00:28:45.605 }, 00:28:45.605 { 00:28:45.605 "name": "BaseBdev3", 00:28:45.605 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:45.605 "is_configured": true, 00:28:45.605 "data_offset": 2048, 00:28:45.605 "data_size": 63488 00:28:45.605 }, 00:28:45.605 { 00:28:45.605 "name": "BaseBdev4", 00:28:45.605 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:45.605 "is_configured": true, 00:28:45.605 "data_offset": 2048, 00:28:45.605 "data_size": 63488 00:28:45.605 } 00:28:45.605 ] 00:28:45.605 }' 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:45.605 [2024-10-01 20:27:40.756574] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:45.605 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:45.606 "name": "raid_bdev1", 00:28:45.606 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:45.606 "strip_size_kb": 0, 00:28:45.606 "state": "online", 00:28:45.606 "raid_level": "raid1", 00:28:45.606 "superblock": true, 00:28:45.606 "num_base_bdevs": 4, 00:28:45.606 "num_base_bdevs_discovered": 2, 00:28:45.606 "num_base_bdevs_operational": 2, 00:28:45.606 "base_bdevs_list": [ 00:28:45.606 { 00:28:45.606 "name": null, 00:28:45.606 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:45.606 "is_configured": false, 00:28:45.606 "data_offset": 0, 00:28:45.606 "data_size": 63488 00:28:45.606 }, 00:28:45.606 { 00:28:45.606 "name": null, 00:28:45.606 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:45.606 "is_configured": false, 00:28:45.606 "data_offset": 2048, 00:28:45.606 "data_size": 63488 00:28:45.606 }, 00:28:45.606 { 00:28:45.606 "name": "BaseBdev3", 00:28:45.606 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:45.606 "is_configured": true, 00:28:45.606 "data_offset": 2048, 00:28:45.606 "data_size": 63488 00:28:45.606 }, 00:28:45.606 { 00:28:45.606 "name": "BaseBdev4", 00:28:45.606 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:45.606 "is_configured": true, 00:28:45.606 "data_offset": 2048, 00:28:45.606 "data_size": 63488 00:28:45.606 } 00:28:45.606 ] 00:28:45.606 }' 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:45.606 20:27:40 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:46.174 20:27:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:28:46.175 20:27:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:46.175 20:27:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:46.175 [2024-10-01 20:27:41.272866] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:46.175 [2024-10-01 20:27:41.273169] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:28:46.175 [2024-10-01 20:27:41.273189] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:28:46.175 [2024-10-01 20:27:41.273263] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:46.175 [2024-10-01 20:27:41.285866] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000037230 00:28:46.175 20:27:41 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:46.175 20:27:41 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:28:46.175 [2024-10-01 20:27:41.288429] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:28:47.112 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:47.112 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:47.112 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:47.112 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:47.112 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:47.112 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:47.112 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:47.112 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:47.112 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:47.112 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:47.112 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:47.112 "name": "raid_bdev1", 00:28:47.112 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:47.112 "strip_size_kb": 0, 00:28:47.112 "state": "online", 00:28:47.113 "raid_level": "raid1", 00:28:47.113 "superblock": true, 00:28:47.113 "num_base_bdevs": 4, 00:28:47.113 "num_base_bdevs_discovered": 3, 00:28:47.113 "num_base_bdevs_operational": 3, 00:28:47.113 "process": { 00:28:47.113 "type": "rebuild", 00:28:47.113 "target": "spare", 00:28:47.113 "progress": { 00:28:47.113 "blocks": 20480, 00:28:47.113 "percent": 32 00:28:47.113 } 00:28:47.113 }, 00:28:47.113 "base_bdevs_list": [ 00:28:47.113 { 00:28:47.113 "name": "spare", 00:28:47.113 "uuid": "a4791e79-ff6a-5a2c-88ad-00ae2225591b", 00:28:47.113 "is_configured": true, 00:28:47.113 "data_offset": 2048, 00:28:47.113 "data_size": 63488 00:28:47.113 }, 00:28:47.113 { 00:28:47.113 "name": null, 00:28:47.113 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:47.113 "is_configured": false, 00:28:47.113 "data_offset": 2048, 00:28:47.113 "data_size": 63488 00:28:47.113 }, 00:28:47.113 { 00:28:47.113 "name": "BaseBdev3", 00:28:47.113 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:47.113 "is_configured": true, 00:28:47.113 "data_offset": 2048, 00:28:47.113 "data_size": 63488 00:28:47.113 }, 00:28:47.113 { 00:28:47.113 "name": "BaseBdev4", 00:28:47.113 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:47.113 "is_configured": true, 00:28:47.113 "data_offset": 2048, 00:28:47.113 "data_size": 63488 00:28:47.113 } 00:28:47.113 ] 00:28:47.113 }' 00:28:47.113 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:47.372 [2024-10-01 20:27:42.450554] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:47.372 [2024-10-01 20:27:42.497777] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:28:47.372 [2024-10-01 20:27:42.498011] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:47.372 [2024-10-01 20:27:42.498045] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:47.372 [2024-10-01 20:27:42.498058] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:47.372 "name": "raid_bdev1", 00:28:47.372 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:47.372 "strip_size_kb": 0, 00:28:47.372 "state": "online", 00:28:47.372 "raid_level": "raid1", 00:28:47.372 "superblock": true, 00:28:47.372 "num_base_bdevs": 4, 00:28:47.372 "num_base_bdevs_discovered": 2, 00:28:47.372 "num_base_bdevs_operational": 2, 00:28:47.372 "base_bdevs_list": [ 00:28:47.372 { 00:28:47.372 "name": null, 00:28:47.372 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:47.372 "is_configured": false, 00:28:47.372 "data_offset": 0, 00:28:47.372 "data_size": 63488 00:28:47.372 }, 00:28:47.372 { 00:28:47.372 "name": null, 00:28:47.372 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:47.372 "is_configured": false, 00:28:47.372 "data_offset": 2048, 00:28:47.372 "data_size": 63488 00:28:47.372 }, 00:28:47.372 { 00:28:47.372 "name": "BaseBdev3", 00:28:47.372 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:47.372 "is_configured": true, 00:28:47.372 "data_offset": 2048, 00:28:47.372 "data_size": 63488 00:28:47.372 }, 00:28:47.372 { 00:28:47.372 "name": "BaseBdev4", 00:28:47.372 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:47.372 "is_configured": true, 00:28:47.372 "data_offset": 2048, 00:28:47.372 "data_size": 63488 00:28:47.372 } 00:28:47.372 ] 00:28:47.372 }' 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:47.372 20:27:42 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:47.939 20:27:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:28:47.939 20:27:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:47.939 20:27:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:47.939 [2024-10-01 20:27:43.047160] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:28:47.939 [2024-10-01 20:27:43.047411] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:47.939 [2024-10-01 20:27:43.047469] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:28:47.939 [2024-10-01 20:27:43.047487] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:47.939 [2024-10-01 20:27:43.048245] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:47.939 [2024-10-01 20:27:43.048269] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:28:47.939 [2024-10-01 20:27:43.048412] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:28:47.939 [2024-10-01 20:27:43.048431] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:28:47.939 [2024-10-01 20:27:43.048447] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:28:47.939 [2024-10-01 20:27:43.048486] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:28:47.939 spare 00:28:47.939 [2024-10-01 20:27:43.060950] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000037300 00:28:47.939 20:27:43 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:47.939 20:27:43 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:28:47.939 [2024-10-01 20:27:43.063583] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:28:48.875 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:28:48.875 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:48.875 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:28:48.875 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:28:48.875 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:48.875 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:48.875 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:48.875 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:48.875 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:48.875 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:48.875 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:48.875 "name": "raid_bdev1", 00:28:48.875 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:48.875 "strip_size_kb": 0, 00:28:48.875 "state": "online", 00:28:48.875 "raid_level": "raid1", 00:28:48.875 "superblock": true, 00:28:48.875 "num_base_bdevs": 4, 00:28:48.875 "num_base_bdevs_discovered": 3, 00:28:48.875 "num_base_bdevs_operational": 3, 00:28:48.875 "process": { 00:28:48.875 "type": "rebuild", 00:28:48.875 "target": "spare", 00:28:48.875 "progress": { 00:28:48.875 "blocks": 20480, 00:28:48.875 "percent": 32 00:28:48.875 } 00:28:48.875 }, 00:28:48.875 "base_bdevs_list": [ 00:28:48.875 { 00:28:48.875 "name": "spare", 00:28:48.875 "uuid": "a4791e79-ff6a-5a2c-88ad-00ae2225591b", 00:28:48.875 "is_configured": true, 00:28:48.875 "data_offset": 2048, 00:28:48.875 "data_size": 63488 00:28:48.875 }, 00:28:48.875 { 00:28:48.875 "name": null, 00:28:48.875 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:48.875 "is_configured": false, 00:28:48.875 "data_offset": 2048, 00:28:48.875 "data_size": 63488 00:28:48.875 }, 00:28:48.875 { 00:28:48.875 "name": "BaseBdev3", 00:28:48.875 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:48.875 "is_configured": true, 00:28:48.875 "data_offset": 2048, 00:28:48.875 "data_size": 63488 00:28:48.875 }, 00:28:48.875 { 00:28:48.875 "name": "BaseBdev4", 00:28:48.875 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:48.875 "is_configured": true, 00:28:48.875 "data_offset": 2048, 00:28:48.875 "data_size": 63488 00:28:48.875 } 00:28:48.875 ] 00:28:48.875 }' 00:28:48.875 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:49.134 [2024-10-01 20:27:44.237514] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:49.134 [2024-10-01 20:27:44.272929] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:28:49.134 [2024-10-01 20:27:44.273177] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:49.134 [2024-10-01 20:27:44.273206] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:28:49.134 [2024-10-01 20:27:44.273222] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:49.134 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:49.134 "name": "raid_bdev1", 00:28:49.134 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:49.134 "strip_size_kb": 0, 00:28:49.134 "state": "online", 00:28:49.134 "raid_level": "raid1", 00:28:49.134 "superblock": true, 00:28:49.134 "num_base_bdevs": 4, 00:28:49.134 "num_base_bdevs_discovered": 2, 00:28:49.134 "num_base_bdevs_operational": 2, 00:28:49.134 "base_bdevs_list": [ 00:28:49.134 { 00:28:49.134 "name": null, 00:28:49.134 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:49.134 "is_configured": false, 00:28:49.134 "data_offset": 0, 00:28:49.135 "data_size": 63488 00:28:49.135 }, 00:28:49.135 { 00:28:49.135 "name": null, 00:28:49.135 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:49.135 "is_configured": false, 00:28:49.135 "data_offset": 2048, 00:28:49.135 "data_size": 63488 00:28:49.135 }, 00:28:49.135 { 00:28:49.135 "name": "BaseBdev3", 00:28:49.135 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:49.135 "is_configured": true, 00:28:49.135 "data_offset": 2048, 00:28:49.135 "data_size": 63488 00:28:49.135 }, 00:28:49.135 { 00:28:49.135 "name": "BaseBdev4", 00:28:49.135 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:49.135 "is_configured": true, 00:28:49.135 "data_offset": 2048, 00:28:49.135 "data_size": 63488 00:28:49.135 } 00:28:49.135 ] 00:28:49.135 }' 00:28:49.135 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:49.135 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:49.702 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:49.702 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:49.702 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:28:49.702 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:28:49.702 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:49.702 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:49.702 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:49.702 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:49.702 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:49.702 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:49.702 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:49.702 "name": "raid_bdev1", 00:28:49.702 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:49.702 "strip_size_kb": 0, 00:28:49.702 "state": "online", 00:28:49.702 "raid_level": "raid1", 00:28:49.702 "superblock": true, 00:28:49.702 "num_base_bdevs": 4, 00:28:49.702 "num_base_bdevs_discovered": 2, 00:28:49.702 "num_base_bdevs_operational": 2, 00:28:49.702 "base_bdevs_list": [ 00:28:49.702 { 00:28:49.702 "name": null, 00:28:49.702 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:49.702 "is_configured": false, 00:28:49.702 "data_offset": 0, 00:28:49.702 "data_size": 63488 00:28:49.702 }, 00:28:49.702 { 00:28:49.702 "name": null, 00:28:49.702 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:49.702 "is_configured": false, 00:28:49.702 "data_offset": 2048, 00:28:49.702 "data_size": 63488 00:28:49.702 }, 00:28:49.702 { 00:28:49.702 "name": "BaseBdev3", 00:28:49.702 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:49.702 "is_configured": true, 00:28:49.702 "data_offset": 2048, 00:28:49.702 "data_size": 63488 00:28:49.702 }, 00:28:49.702 { 00:28:49.702 "name": "BaseBdev4", 00:28:49.702 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:49.702 "is_configured": true, 00:28:49.702 "data_offset": 2048, 00:28:49.702 "data_size": 63488 00:28:49.702 } 00:28:49.702 ] 00:28:49.702 }' 00:28:49.703 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:49.703 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:28:49.703 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:49.961 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:28:49.961 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:28:49.961 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:49.961 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:49.961 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:49.961 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:28:49.961 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:49.961 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:49.961 [2024-10-01 20:27:44.993016] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:28:49.961 [2024-10-01 20:27:44.993273] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:28:49.962 [2024-10-01 20:27:44.993358] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000cc80 00:28:49.962 [2024-10-01 20:27:44.993379] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:28:49.962 [2024-10-01 20:27:44.994036] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:28:49.962 [2024-10-01 20:27:44.994100] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:28:49.962 [2024-10-01 20:27:44.994203] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:28:49.962 [2024-10-01 20:27:44.994228] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:28:49.962 [2024-10-01 20:27:44.994240] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:28:49.962 [2024-10-01 20:27:44.994255] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:28:49.962 BaseBdev1 00:28:49.962 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:49.962 20:27:44 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:28:50.897 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:50.897 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:50.897 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:50.897 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:50.897 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:50.897 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:50.897 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:50.897 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:50.897 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:50.897 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:50.898 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:50.898 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:50.898 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:50.898 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:50.898 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:50.898 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:50.898 "name": "raid_bdev1", 00:28:50.898 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:50.898 "strip_size_kb": 0, 00:28:50.898 "state": "online", 00:28:50.898 "raid_level": "raid1", 00:28:50.898 "superblock": true, 00:28:50.898 "num_base_bdevs": 4, 00:28:50.898 "num_base_bdevs_discovered": 2, 00:28:50.898 "num_base_bdevs_operational": 2, 00:28:50.898 "base_bdevs_list": [ 00:28:50.898 { 00:28:50.898 "name": null, 00:28:50.898 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:50.898 "is_configured": false, 00:28:50.898 "data_offset": 0, 00:28:50.898 "data_size": 63488 00:28:50.898 }, 00:28:50.898 { 00:28:50.898 "name": null, 00:28:50.898 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:50.898 "is_configured": false, 00:28:50.898 "data_offset": 2048, 00:28:50.898 "data_size": 63488 00:28:50.898 }, 00:28:50.898 { 00:28:50.898 "name": "BaseBdev3", 00:28:50.898 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:50.898 "is_configured": true, 00:28:50.898 "data_offset": 2048, 00:28:50.898 "data_size": 63488 00:28:50.898 }, 00:28:50.898 { 00:28:50.898 "name": "BaseBdev4", 00:28:50.898 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:50.898 "is_configured": true, 00:28:50.898 "data_offset": 2048, 00:28:50.898 "data_size": 63488 00:28:50.898 } 00:28:50.898 ] 00:28:50.898 }' 00:28:50.898 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:50.898 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:51.497 "name": "raid_bdev1", 00:28:51.497 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:51.497 "strip_size_kb": 0, 00:28:51.497 "state": "online", 00:28:51.497 "raid_level": "raid1", 00:28:51.497 "superblock": true, 00:28:51.497 "num_base_bdevs": 4, 00:28:51.497 "num_base_bdevs_discovered": 2, 00:28:51.497 "num_base_bdevs_operational": 2, 00:28:51.497 "base_bdevs_list": [ 00:28:51.497 { 00:28:51.497 "name": null, 00:28:51.497 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:51.497 "is_configured": false, 00:28:51.497 "data_offset": 0, 00:28:51.497 "data_size": 63488 00:28:51.497 }, 00:28:51.497 { 00:28:51.497 "name": null, 00:28:51.497 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:51.497 "is_configured": false, 00:28:51.497 "data_offset": 2048, 00:28:51.497 "data_size": 63488 00:28:51.497 }, 00:28:51.497 { 00:28:51.497 "name": "BaseBdev3", 00:28:51.497 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:51.497 "is_configured": true, 00:28:51.497 "data_offset": 2048, 00:28:51.497 "data_size": 63488 00:28:51.497 }, 00:28:51.497 { 00:28:51.497 "name": "BaseBdev4", 00:28:51.497 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:51.497 "is_configured": true, 00:28:51.497 "data_offset": 2048, 00:28:51.497 "data_size": 63488 00:28:51.497 } 00:28:51.497 ] 00:28:51.497 }' 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@650 -- # local es=0 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:51.497 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:51.497 [2024-10-01 20:27:46.698064] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:51.497 [2024-10-01 20:27:46.698458] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:28:51.497 [2024-10-01 20:27:46.698488] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:28:51.497 request: 00:28:51.497 { 00:28:51.497 "base_bdev": "BaseBdev1", 00:28:51.497 "raid_bdev": "raid_bdev1", 00:28:51.497 "method": "bdev_raid_add_base_bdev", 00:28:51.497 "req_id": 1 00:28:51.497 } 00:28:51.497 Got JSON-RPC error response 00:28:51.497 response: 00:28:51.498 { 00:28:51.498 "code": -22, 00:28:51.498 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:28:51.498 } 00:28:51.498 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:28:51.498 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # es=1 00:28:51.498 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:28:51.498 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:28:51.498 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:28:51.498 20:27:46 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:28:52.874 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:28:52.874 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:28:52.874 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:52.874 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:28:52.874 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:28:52.874 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:28:52.874 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:52.874 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:52.874 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:52.874 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:52.874 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:52.874 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:52.874 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:52.874 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:52.875 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:52.875 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:52.875 "name": "raid_bdev1", 00:28:52.875 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:52.875 "strip_size_kb": 0, 00:28:52.875 "state": "online", 00:28:52.875 "raid_level": "raid1", 00:28:52.875 "superblock": true, 00:28:52.875 "num_base_bdevs": 4, 00:28:52.875 "num_base_bdevs_discovered": 2, 00:28:52.875 "num_base_bdevs_operational": 2, 00:28:52.875 "base_bdevs_list": [ 00:28:52.875 { 00:28:52.875 "name": null, 00:28:52.875 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:52.875 "is_configured": false, 00:28:52.875 "data_offset": 0, 00:28:52.875 "data_size": 63488 00:28:52.875 }, 00:28:52.875 { 00:28:52.875 "name": null, 00:28:52.875 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:52.875 "is_configured": false, 00:28:52.875 "data_offset": 2048, 00:28:52.875 "data_size": 63488 00:28:52.875 }, 00:28:52.875 { 00:28:52.875 "name": "BaseBdev3", 00:28:52.875 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:52.875 "is_configured": true, 00:28:52.875 "data_offset": 2048, 00:28:52.875 "data_size": 63488 00:28:52.875 }, 00:28:52.875 { 00:28:52.875 "name": "BaseBdev4", 00:28:52.875 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:52.875 "is_configured": true, 00:28:52.875 "data_offset": 2048, 00:28:52.875 "data_size": 63488 00:28:52.875 } 00:28:52.875 ] 00:28:52.875 }' 00:28:52.875 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:52.875 20:27:47 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:53.133 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:28:53.134 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:28:53.134 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:28:53.134 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:28:53.134 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:28:53.134 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:53.134 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:28:53.134 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:53.134 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:53.134 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:53.134 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:28:53.134 "name": "raid_bdev1", 00:28:53.134 "uuid": "e02ccd86-cabf-4f91-8133-37fbe7d75b4f", 00:28:53.134 "strip_size_kb": 0, 00:28:53.134 "state": "online", 00:28:53.134 "raid_level": "raid1", 00:28:53.134 "superblock": true, 00:28:53.134 "num_base_bdevs": 4, 00:28:53.134 "num_base_bdevs_discovered": 2, 00:28:53.134 "num_base_bdevs_operational": 2, 00:28:53.134 "base_bdevs_list": [ 00:28:53.134 { 00:28:53.134 "name": null, 00:28:53.134 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:53.134 "is_configured": false, 00:28:53.134 "data_offset": 0, 00:28:53.134 "data_size": 63488 00:28:53.134 }, 00:28:53.134 { 00:28:53.134 "name": null, 00:28:53.134 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:53.134 "is_configured": false, 00:28:53.134 "data_offset": 2048, 00:28:53.134 "data_size": 63488 00:28:53.134 }, 00:28:53.134 { 00:28:53.134 "name": "BaseBdev3", 00:28:53.134 "uuid": "c2e23be3-51f8-5a45-8227-5e58c1ef07c8", 00:28:53.134 "is_configured": true, 00:28:53.134 "data_offset": 2048, 00:28:53.134 "data_size": 63488 00:28:53.134 }, 00:28:53.134 { 00:28:53.134 "name": "BaseBdev4", 00:28:53.134 "uuid": "a0c41bc5-40a4-580f-9ca9-0fa3b3c22c6c", 00:28:53.134 "is_configured": true, 00:28:53.134 "data_offset": 2048, 00:28:53.134 "data_size": 63488 00:28:53.134 } 00:28:53.134 ] 00:28:53.134 }' 00:28:53.134 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:28:53.134 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:28:53.134 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:28:53.393 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:28:53.393 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 80349 00:28:53.393 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@950 -- # '[' -z 80349 ']' 00:28:53.393 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # kill -0 80349 00:28:53.393 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@955 -- # uname 00:28:53.393 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:28:53.393 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 80349 00:28:53.393 killing process with pid 80349 00:28:53.393 Received shutdown signal, test time was about 18.945965 seconds 00:28:53.393 00:28:53.393 Latency(us) 00:28:53.393 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:53.393 =================================================================================================================== 00:28:53.393 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:28:53.393 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:28:53.393 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:28:53.393 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@968 -- # echo 'killing process with pid 80349' 00:28:53.393 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@969 -- # kill 80349 00:28:53.394 [2024-10-01 20:27:48.440707] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:28:53.394 20:27:48 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@974 -- # wait 80349 00:28:53.394 [2024-10-01 20:27:48.440935] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:28:53.394 [2024-10-01 20:27:48.441039] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:28:53.394 [2024-10-01 20:27:48.441057] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:28:53.653 [2024-10-01 20:27:48.792001] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:28:56.188 20:27:50 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:28:56.188 00:28:56.188 real 0m23.527s 00:28:56.188 user 0m31.545s 00:28:56.188 sys 0m2.493s 00:28:56.188 20:27:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:28:56.188 20:27:50 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:28:56.188 ************************************ 00:28:56.188 END TEST raid_rebuild_test_sb_io 00:28:56.188 ************************************ 00:28:56.188 20:27:50 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:28:56.188 20:27:50 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 3 false 00:28:56.188 20:27:50 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:28:56.188 20:27:50 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:28:56.188 20:27:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:28:56.188 ************************************ 00:28:56.188 START TEST raid5f_state_function_test 00:28:56.188 ************************************ 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid5f 3 false 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:28:56.188 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:28:56.189 Process raid pid: 81088 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=81088 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 81088' 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 81088 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 81088 ']' 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:28:56.189 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:28:56.189 20:27:50 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:56.189 [2024-10-01 20:27:51.032386] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:28:56.189 [2024-10-01 20:27:51.032602] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:28:56.189 [2024-10-01 20:27:51.211593] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:56.448 [2024-10-01 20:27:51.481100] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:28:56.706 [2024-10-01 20:27:51.722878] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:56.706 [2024-10-01 20:27:51.722925] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:56.964 [2024-10-01 20:27:52.177049] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:56.964 [2024-10-01 20:27:52.177248] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:56.964 [2024-10-01 20:27:52.177450] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:56.964 [2024-10-01 20:27:52.177488] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:56.964 [2024-10-01 20:27:52.177501] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:28:56.964 [2024-10-01 20:27:52.177517] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:56.964 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:57.223 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:57.223 "name": "Existed_Raid", 00:28:57.223 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:57.223 "strip_size_kb": 64, 00:28:57.223 "state": "configuring", 00:28:57.223 "raid_level": "raid5f", 00:28:57.223 "superblock": false, 00:28:57.223 "num_base_bdevs": 3, 00:28:57.223 "num_base_bdevs_discovered": 0, 00:28:57.223 "num_base_bdevs_operational": 3, 00:28:57.223 "base_bdevs_list": [ 00:28:57.223 { 00:28:57.223 "name": "BaseBdev1", 00:28:57.223 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:57.223 "is_configured": false, 00:28:57.223 "data_offset": 0, 00:28:57.223 "data_size": 0 00:28:57.223 }, 00:28:57.223 { 00:28:57.223 "name": "BaseBdev2", 00:28:57.223 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:57.223 "is_configured": false, 00:28:57.223 "data_offset": 0, 00:28:57.223 "data_size": 0 00:28:57.223 }, 00:28:57.223 { 00:28:57.223 "name": "BaseBdev3", 00:28:57.223 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:57.223 "is_configured": false, 00:28:57.223 "data_offset": 0, 00:28:57.223 "data_size": 0 00:28:57.223 } 00:28:57.223 ] 00:28:57.223 }' 00:28:57.223 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:57.223 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.482 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:28:57.482 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:57.482 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.482 [2024-10-01 20:27:52.713128] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:28:57.482 [2024-10-01 20:27:52.713173] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:28:57.482 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:57.482 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:28:57.482 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:57.482 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.482 [2024-10-01 20:27:52.721095] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:28:57.482 [2024-10-01 20:27:52.721280] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:28:57.482 [2024-10-01 20:27:52.721425] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:57.482 [2024-10-01 20:27:52.721488] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:57.482 [2024-10-01 20:27:52.721636] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:28:57.482 [2024-10-01 20:27:52.721707] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:28:57.482 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:57.482 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:28:57.482 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:57.482 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.750 [2024-10-01 20:27:52.773476] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:57.750 BaseBdev1 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.750 [ 00:28:57.750 { 00:28:57.750 "name": "BaseBdev1", 00:28:57.750 "aliases": [ 00:28:57.750 "77226df1-ca20-4888-9a71-5b6ab6c989ec" 00:28:57.750 ], 00:28:57.750 "product_name": "Malloc disk", 00:28:57.750 "block_size": 512, 00:28:57.750 "num_blocks": 65536, 00:28:57.750 "uuid": "77226df1-ca20-4888-9a71-5b6ab6c989ec", 00:28:57.750 "assigned_rate_limits": { 00:28:57.750 "rw_ios_per_sec": 0, 00:28:57.750 "rw_mbytes_per_sec": 0, 00:28:57.750 "r_mbytes_per_sec": 0, 00:28:57.750 "w_mbytes_per_sec": 0 00:28:57.750 }, 00:28:57.750 "claimed": true, 00:28:57.750 "claim_type": "exclusive_write", 00:28:57.750 "zoned": false, 00:28:57.750 "supported_io_types": { 00:28:57.750 "read": true, 00:28:57.750 "write": true, 00:28:57.750 "unmap": true, 00:28:57.750 "flush": true, 00:28:57.750 "reset": true, 00:28:57.750 "nvme_admin": false, 00:28:57.750 "nvme_io": false, 00:28:57.750 "nvme_io_md": false, 00:28:57.750 "write_zeroes": true, 00:28:57.750 "zcopy": true, 00:28:57.750 "get_zone_info": false, 00:28:57.750 "zone_management": false, 00:28:57.750 "zone_append": false, 00:28:57.750 "compare": false, 00:28:57.750 "compare_and_write": false, 00:28:57.750 "abort": true, 00:28:57.750 "seek_hole": false, 00:28:57.750 "seek_data": false, 00:28:57.750 "copy": true, 00:28:57.750 "nvme_iov_md": false 00:28:57.750 }, 00:28:57.750 "memory_domains": [ 00:28:57.750 { 00:28:57.750 "dma_device_id": "system", 00:28:57.750 "dma_device_type": 1 00:28:57.750 }, 00:28:57.750 { 00:28:57.750 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:57.750 "dma_device_type": 2 00:28:57.750 } 00:28:57.750 ], 00:28:57.750 "driver_specific": {} 00:28:57.750 } 00:28:57.750 ] 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:57.750 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:57.750 "name": "Existed_Raid", 00:28:57.750 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:57.750 "strip_size_kb": 64, 00:28:57.750 "state": "configuring", 00:28:57.750 "raid_level": "raid5f", 00:28:57.750 "superblock": false, 00:28:57.750 "num_base_bdevs": 3, 00:28:57.750 "num_base_bdevs_discovered": 1, 00:28:57.750 "num_base_bdevs_operational": 3, 00:28:57.750 "base_bdevs_list": [ 00:28:57.750 { 00:28:57.750 "name": "BaseBdev1", 00:28:57.750 "uuid": "77226df1-ca20-4888-9a71-5b6ab6c989ec", 00:28:57.750 "is_configured": true, 00:28:57.750 "data_offset": 0, 00:28:57.750 "data_size": 65536 00:28:57.750 }, 00:28:57.750 { 00:28:57.750 "name": "BaseBdev2", 00:28:57.750 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:57.750 "is_configured": false, 00:28:57.750 "data_offset": 0, 00:28:57.750 "data_size": 0 00:28:57.750 }, 00:28:57.750 { 00:28:57.750 "name": "BaseBdev3", 00:28:57.750 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:57.750 "is_configured": false, 00:28:57.750 "data_offset": 0, 00:28:57.750 "data_size": 0 00:28:57.750 } 00:28:57.750 ] 00:28:57.750 }' 00:28:57.751 20:27:52 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:57.751 20:27:52 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.332 [2024-10-01 20:27:53.325684] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:28:58.332 [2024-10-01 20:27:53.325801] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.332 [2024-10-01 20:27:53.333700] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:28:58.332 [2024-10-01 20:27:53.336821] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:28:58.332 [2024-10-01 20:27:53.336999] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:28:58.332 [2024-10-01 20:27:53.337122] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:28:58.332 [2024-10-01 20:27:53.337256] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:58.332 "name": "Existed_Raid", 00:28:58.332 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:58.332 "strip_size_kb": 64, 00:28:58.332 "state": "configuring", 00:28:58.332 "raid_level": "raid5f", 00:28:58.332 "superblock": false, 00:28:58.332 "num_base_bdevs": 3, 00:28:58.332 "num_base_bdevs_discovered": 1, 00:28:58.332 "num_base_bdevs_operational": 3, 00:28:58.332 "base_bdevs_list": [ 00:28:58.332 { 00:28:58.332 "name": "BaseBdev1", 00:28:58.332 "uuid": "77226df1-ca20-4888-9a71-5b6ab6c989ec", 00:28:58.332 "is_configured": true, 00:28:58.332 "data_offset": 0, 00:28:58.332 "data_size": 65536 00:28:58.332 }, 00:28:58.332 { 00:28:58.332 "name": "BaseBdev2", 00:28:58.332 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:58.332 "is_configured": false, 00:28:58.332 "data_offset": 0, 00:28:58.332 "data_size": 0 00:28:58.332 }, 00:28:58.332 { 00:28:58.332 "name": "BaseBdev3", 00:28:58.332 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:58.332 "is_configured": false, 00:28:58.332 "data_offset": 0, 00:28:58.332 "data_size": 0 00:28:58.332 } 00:28:58.332 ] 00:28:58.332 }' 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:58.332 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.900 [2024-10-01 20:27:53.919732] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:28:58.900 BaseBdev2 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.900 [ 00:28:58.900 { 00:28:58.900 "name": "BaseBdev2", 00:28:58.900 "aliases": [ 00:28:58.900 "c850ea7c-2b1e-46b6-ae79-bdab6b13104e" 00:28:58.900 ], 00:28:58.900 "product_name": "Malloc disk", 00:28:58.900 "block_size": 512, 00:28:58.900 "num_blocks": 65536, 00:28:58.900 "uuid": "c850ea7c-2b1e-46b6-ae79-bdab6b13104e", 00:28:58.900 "assigned_rate_limits": { 00:28:58.900 "rw_ios_per_sec": 0, 00:28:58.900 "rw_mbytes_per_sec": 0, 00:28:58.900 "r_mbytes_per_sec": 0, 00:28:58.900 "w_mbytes_per_sec": 0 00:28:58.900 }, 00:28:58.900 "claimed": true, 00:28:58.900 "claim_type": "exclusive_write", 00:28:58.900 "zoned": false, 00:28:58.900 "supported_io_types": { 00:28:58.900 "read": true, 00:28:58.900 "write": true, 00:28:58.900 "unmap": true, 00:28:58.900 "flush": true, 00:28:58.900 "reset": true, 00:28:58.900 "nvme_admin": false, 00:28:58.900 "nvme_io": false, 00:28:58.900 "nvme_io_md": false, 00:28:58.900 "write_zeroes": true, 00:28:58.900 "zcopy": true, 00:28:58.900 "get_zone_info": false, 00:28:58.900 "zone_management": false, 00:28:58.900 "zone_append": false, 00:28:58.900 "compare": false, 00:28:58.900 "compare_and_write": false, 00:28:58.900 "abort": true, 00:28:58.900 "seek_hole": false, 00:28:58.900 "seek_data": false, 00:28:58.900 "copy": true, 00:28:58.900 "nvme_iov_md": false 00:28:58.900 }, 00:28:58.900 "memory_domains": [ 00:28:58.900 { 00:28:58.900 "dma_device_id": "system", 00:28:58.900 "dma_device_type": 1 00:28:58.900 }, 00:28:58.900 { 00:28:58.900 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:58.900 "dma_device_type": 2 00:28:58.900 } 00:28:58.900 ], 00:28:58.900 "driver_specific": {} 00:28:58.900 } 00:28:58.900 ] 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:58.900 20:27:53 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:58.900 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:58.900 "name": "Existed_Raid", 00:28:58.900 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:58.900 "strip_size_kb": 64, 00:28:58.900 "state": "configuring", 00:28:58.900 "raid_level": "raid5f", 00:28:58.900 "superblock": false, 00:28:58.900 "num_base_bdevs": 3, 00:28:58.900 "num_base_bdevs_discovered": 2, 00:28:58.900 "num_base_bdevs_operational": 3, 00:28:58.900 "base_bdevs_list": [ 00:28:58.900 { 00:28:58.900 "name": "BaseBdev1", 00:28:58.900 "uuid": "77226df1-ca20-4888-9a71-5b6ab6c989ec", 00:28:58.900 "is_configured": true, 00:28:58.900 "data_offset": 0, 00:28:58.900 "data_size": 65536 00:28:58.900 }, 00:28:58.900 { 00:28:58.900 "name": "BaseBdev2", 00:28:58.900 "uuid": "c850ea7c-2b1e-46b6-ae79-bdab6b13104e", 00:28:58.900 "is_configured": true, 00:28:58.900 "data_offset": 0, 00:28:58.900 "data_size": 65536 00:28:58.900 }, 00:28:58.900 { 00:28:58.900 "name": "BaseBdev3", 00:28:58.900 "uuid": "00000000-0000-0000-0000-000000000000", 00:28:58.900 "is_configured": false, 00:28:58.900 "data_offset": 0, 00:28:58.900 "data_size": 0 00:28:58.900 } 00:28:58.900 ] 00:28:58.900 }' 00:28:58.900 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:58.900 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:59.468 [2024-10-01 20:27:54.520266] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:28:59.468 [2024-10-01 20:27:54.520553] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:28:59.468 [2024-10-01 20:27:54.520592] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:28:59.468 [2024-10-01 20:27:54.520980] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:28:59.468 BaseBdev3 00:28:59.468 [2024-10-01 20:27:54.526726] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:28:59.468 [2024-10-01 20:27:54.526754] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:28:59.468 [2024-10-01 20:27:54.527135] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:59.468 [ 00:28:59.468 { 00:28:59.468 "name": "BaseBdev3", 00:28:59.468 "aliases": [ 00:28:59.468 "640df568-3a7c-4260-968f-15c1bfdd71fc" 00:28:59.468 ], 00:28:59.468 "product_name": "Malloc disk", 00:28:59.468 "block_size": 512, 00:28:59.468 "num_blocks": 65536, 00:28:59.468 "uuid": "640df568-3a7c-4260-968f-15c1bfdd71fc", 00:28:59.468 "assigned_rate_limits": { 00:28:59.468 "rw_ios_per_sec": 0, 00:28:59.468 "rw_mbytes_per_sec": 0, 00:28:59.468 "r_mbytes_per_sec": 0, 00:28:59.468 "w_mbytes_per_sec": 0 00:28:59.468 }, 00:28:59.468 "claimed": true, 00:28:59.468 "claim_type": "exclusive_write", 00:28:59.468 "zoned": false, 00:28:59.468 "supported_io_types": { 00:28:59.468 "read": true, 00:28:59.468 "write": true, 00:28:59.468 "unmap": true, 00:28:59.468 "flush": true, 00:28:59.468 "reset": true, 00:28:59.468 "nvme_admin": false, 00:28:59.468 "nvme_io": false, 00:28:59.468 "nvme_io_md": false, 00:28:59.468 "write_zeroes": true, 00:28:59.468 "zcopy": true, 00:28:59.468 "get_zone_info": false, 00:28:59.468 "zone_management": false, 00:28:59.468 "zone_append": false, 00:28:59.468 "compare": false, 00:28:59.468 "compare_and_write": false, 00:28:59.468 "abort": true, 00:28:59.468 "seek_hole": false, 00:28:59.468 "seek_data": false, 00:28:59.468 "copy": true, 00:28:59.468 "nvme_iov_md": false 00:28:59.468 }, 00:28:59.468 "memory_domains": [ 00:28:59.468 { 00:28:59.468 "dma_device_id": "system", 00:28:59.468 "dma_device_type": 1 00:28:59.468 }, 00:28:59.468 { 00:28:59.468 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:28:59.468 "dma_device_type": 2 00:28:59.468 } 00:28:59.468 ], 00:28:59.468 "driver_specific": {} 00:28:59.468 } 00:28:59.468 ] 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:28:59.468 "name": "Existed_Raid", 00:28:59.468 "uuid": "ae21a2e7-2e52-41f1-8c57-96670f63b05d", 00:28:59.468 "strip_size_kb": 64, 00:28:59.468 "state": "online", 00:28:59.468 "raid_level": "raid5f", 00:28:59.468 "superblock": false, 00:28:59.468 "num_base_bdevs": 3, 00:28:59.468 "num_base_bdevs_discovered": 3, 00:28:59.468 "num_base_bdevs_operational": 3, 00:28:59.468 "base_bdevs_list": [ 00:28:59.468 { 00:28:59.468 "name": "BaseBdev1", 00:28:59.468 "uuid": "77226df1-ca20-4888-9a71-5b6ab6c989ec", 00:28:59.468 "is_configured": true, 00:28:59.468 "data_offset": 0, 00:28:59.468 "data_size": 65536 00:28:59.468 }, 00:28:59.468 { 00:28:59.468 "name": "BaseBdev2", 00:28:59.468 "uuid": "c850ea7c-2b1e-46b6-ae79-bdab6b13104e", 00:28:59.468 "is_configured": true, 00:28:59.468 "data_offset": 0, 00:28:59.468 "data_size": 65536 00:28:59.468 }, 00:28:59.468 { 00:28:59.468 "name": "BaseBdev3", 00:28:59.468 "uuid": "640df568-3a7c-4260-968f-15c1bfdd71fc", 00:28:59.468 "is_configured": true, 00:28:59.468 "data_offset": 0, 00:28:59.468 "data_size": 65536 00:28:59.468 } 00:28:59.468 ] 00:28:59.468 }' 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:28:59.468 20:27:54 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.036 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:00.037 [2024-10-01 20:27:55.109417] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:00.037 "name": "Existed_Raid", 00:29:00.037 "aliases": [ 00:29:00.037 "ae21a2e7-2e52-41f1-8c57-96670f63b05d" 00:29:00.037 ], 00:29:00.037 "product_name": "Raid Volume", 00:29:00.037 "block_size": 512, 00:29:00.037 "num_blocks": 131072, 00:29:00.037 "uuid": "ae21a2e7-2e52-41f1-8c57-96670f63b05d", 00:29:00.037 "assigned_rate_limits": { 00:29:00.037 "rw_ios_per_sec": 0, 00:29:00.037 "rw_mbytes_per_sec": 0, 00:29:00.037 "r_mbytes_per_sec": 0, 00:29:00.037 "w_mbytes_per_sec": 0 00:29:00.037 }, 00:29:00.037 "claimed": false, 00:29:00.037 "zoned": false, 00:29:00.037 "supported_io_types": { 00:29:00.037 "read": true, 00:29:00.037 "write": true, 00:29:00.037 "unmap": false, 00:29:00.037 "flush": false, 00:29:00.037 "reset": true, 00:29:00.037 "nvme_admin": false, 00:29:00.037 "nvme_io": false, 00:29:00.037 "nvme_io_md": false, 00:29:00.037 "write_zeroes": true, 00:29:00.037 "zcopy": false, 00:29:00.037 "get_zone_info": false, 00:29:00.037 "zone_management": false, 00:29:00.037 "zone_append": false, 00:29:00.037 "compare": false, 00:29:00.037 "compare_and_write": false, 00:29:00.037 "abort": false, 00:29:00.037 "seek_hole": false, 00:29:00.037 "seek_data": false, 00:29:00.037 "copy": false, 00:29:00.037 "nvme_iov_md": false 00:29:00.037 }, 00:29:00.037 "driver_specific": { 00:29:00.037 "raid": { 00:29:00.037 "uuid": "ae21a2e7-2e52-41f1-8c57-96670f63b05d", 00:29:00.037 "strip_size_kb": 64, 00:29:00.037 "state": "online", 00:29:00.037 "raid_level": "raid5f", 00:29:00.037 "superblock": false, 00:29:00.037 "num_base_bdevs": 3, 00:29:00.037 "num_base_bdevs_discovered": 3, 00:29:00.037 "num_base_bdevs_operational": 3, 00:29:00.037 "base_bdevs_list": [ 00:29:00.037 { 00:29:00.037 "name": "BaseBdev1", 00:29:00.037 "uuid": "77226df1-ca20-4888-9a71-5b6ab6c989ec", 00:29:00.037 "is_configured": true, 00:29:00.037 "data_offset": 0, 00:29:00.037 "data_size": 65536 00:29:00.037 }, 00:29:00.037 { 00:29:00.037 "name": "BaseBdev2", 00:29:00.037 "uuid": "c850ea7c-2b1e-46b6-ae79-bdab6b13104e", 00:29:00.037 "is_configured": true, 00:29:00.037 "data_offset": 0, 00:29:00.037 "data_size": 65536 00:29:00.037 }, 00:29:00.037 { 00:29:00.037 "name": "BaseBdev3", 00:29:00.037 "uuid": "640df568-3a7c-4260-968f-15c1bfdd71fc", 00:29:00.037 "is_configured": true, 00:29:00.037 "data_offset": 0, 00:29:00.037 "data_size": 65536 00:29:00.037 } 00:29:00.037 ] 00:29:00.037 } 00:29:00.037 } 00:29:00.037 }' 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:29:00.037 BaseBdev2 00:29:00.037 BaseBdev3' 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:00.037 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.295 [2024-10-01 20:27:55.433245] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:00.295 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.554 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:00.554 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:00.554 "name": "Existed_Raid", 00:29:00.554 "uuid": "ae21a2e7-2e52-41f1-8c57-96670f63b05d", 00:29:00.554 "strip_size_kb": 64, 00:29:00.554 "state": "online", 00:29:00.554 "raid_level": "raid5f", 00:29:00.554 "superblock": false, 00:29:00.554 "num_base_bdevs": 3, 00:29:00.554 "num_base_bdevs_discovered": 2, 00:29:00.554 "num_base_bdevs_operational": 2, 00:29:00.554 "base_bdevs_list": [ 00:29:00.555 { 00:29:00.555 "name": null, 00:29:00.555 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:00.555 "is_configured": false, 00:29:00.555 "data_offset": 0, 00:29:00.555 "data_size": 65536 00:29:00.555 }, 00:29:00.555 { 00:29:00.555 "name": "BaseBdev2", 00:29:00.555 "uuid": "c850ea7c-2b1e-46b6-ae79-bdab6b13104e", 00:29:00.555 "is_configured": true, 00:29:00.555 "data_offset": 0, 00:29:00.555 "data_size": 65536 00:29:00.555 }, 00:29:00.555 { 00:29:00.555 "name": "BaseBdev3", 00:29:00.555 "uuid": "640df568-3a7c-4260-968f-15c1bfdd71fc", 00:29:00.555 "is_configured": true, 00:29:00.555 "data_offset": 0, 00:29:00.555 "data_size": 65536 00:29:00.555 } 00:29:00.555 ] 00:29:00.555 }' 00:29:00.555 20:27:55 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:00.555 20:27:55 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.813 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:29:00.813 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:00.813 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:00.813 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:00.813 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:00.813 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:29:00.813 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:01.071 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:29:01.071 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:29:01.072 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:29:01.072 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:01.072 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.072 [2024-10-01 20:27:56.098394] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:29:01.072 [2024-10-01 20:27:56.098665] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:01.072 [2024-10-01 20:27:56.195509] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:01.072 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:01.072 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:29:01.072 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:01.072 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:01.072 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:29:01.072 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:01.072 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.072 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:01.072 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:29:01.072 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:29:01.072 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:29:01.072 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:01.072 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.072 [2024-10-01 20:27:56.255533] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:29:01.072 [2024-10-01 20:27:56.255742] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.330 BaseBdev2 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.330 [ 00:29:01.330 { 00:29:01.330 "name": "BaseBdev2", 00:29:01.330 "aliases": [ 00:29:01.330 "dc746513-7297-412c-8857-dbad6fd7e130" 00:29:01.330 ], 00:29:01.330 "product_name": "Malloc disk", 00:29:01.330 "block_size": 512, 00:29:01.330 "num_blocks": 65536, 00:29:01.330 "uuid": "dc746513-7297-412c-8857-dbad6fd7e130", 00:29:01.330 "assigned_rate_limits": { 00:29:01.330 "rw_ios_per_sec": 0, 00:29:01.330 "rw_mbytes_per_sec": 0, 00:29:01.330 "r_mbytes_per_sec": 0, 00:29:01.330 "w_mbytes_per_sec": 0 00:29:01.330 }, 00:29:01.330 "claimed": false, 00:29:01.330 "zoned": false, 00:29:01.330 "supported_io_types": { 00:29:01.330 "read": true, 00:29:01.330 "write": true, 00:29:01.330 "unmap": true, 00:29:01.330 "flush": true, 00:29:01.330 "reset": true, 00:29:01.330 "nvme_admin": false, 00:29:01.330 "nvme_io": false, 00:29:01.330 "nvme_io_md": false, 00:29:01.330 "write_zeroes": true, 00:29:01.330 "zcopy": true, 00:29:01.330 "get_zone_info": false, 00:29:01.330 "zone_management": false, 00:29:01.330 "zone_append": false, 00:29:01.330 "compare": false, 00:29:01.330 "compare_and_write": false, 00:29:01.330 "abort": true, 00:29:01.330 "seek_hole": false, 00:29:01.330 "seek_data": false, 00:29:01.330 "copy": true, 00:29:01.330 "nvme_iov_md": false 00:29:01.330 }, 00:29:01.330 "memory_domains": [ 00:29:01.330 { 00:29:01.330 "dma_device_id": "system", 00:29:01.330 "dma_device_type": 1 00:29:01.330 }, 00:29:01.330 { 00:29:01.330 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:01.330 "dma_device_type": 2 00:29:01.330 } 00:29:01.330 ], 00:29:01.330 "driver_specific": {} 00:29:01.330 } 00:29:01.330 ] 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.330 BaseBdev3 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.330 [ 00:29:01.330 { 00:29:01.330 "name": "BaseBdev3", 00:29:01.330 "aliases": [ 00:29:01.330 "0a6ba6d1-0664-49ed-a743-f3e9e704f553" 00:29:01.330 ], 00:29:01.330 "product_name": "Malloc disk", 00:29:01.330 "block_size": 512, 00:29:01.330 "num_blocks": 65536, 00:29:01.330 "uuid": "0a6ba6d1-0664-49ed-a743-f3e9e704f553", 00:29:01.330 "assigned_rate_limits": { 00:29:01.330 "rw_ios_per_sec": 0, 00:29:01.330 "rw_mbytes_per_sec": 0, 00:29:01.330 "r_mbytes_per_sec": 0, 00:29:01.330 "w_mbytes_per_sec": 0 00:29:01.330 }, 00:29:01.330 "claimed": false, 00:29:01.330 "zoned": false, 00:29:01.330 "supported_io_types": { 00:29:01.330 "read": true, 00:29:01.330 "write": true, 00:29:01.330 "unmap": true, 00:29:01.330 "flush": true, 00:29:01.330 "reset": true, 00:29:01.330 "nvme_admin": false, 00:29:01.330 "nvme_io": false, 00:29:01.330 "nvme_io_md": false, 00:29:01.330 "write_zeroes": true, 00:29:01.330 "zcopy": true, 00:29:01.330 "get_zone_info": false, 00:29:01.330 "zone_management": false, 00:29:01.330 "zone_append": false, 00:29:01.330 "compare": false, 00:29:01.330 "compare_and_write": false, 00:29:01.330 "abort": true, 00:29:01.330 "seek_hole": false, 00:29:01.330 "seek_data": false, 00:29:01.330 "copy": true, 00:29:01.330 "nvme_iov_md": false 00:29:01.330 }, 00:29:01.330 "memory_domains": [ 00:29:01.330 { 00:29:01.330 "dma_device_id": "system", 00:29:01.330 "dma_device_type": 1 00:29:01.330 }, 00:29:01.330 { 00:29:01.330 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:01.330 "dma_device_type": 2 00:29:01.330 } 00:29:01.330 ], 00:29:01.330 "driver_specific": {} 00:29:01.330 } 00:29:01.330 ] 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:01.330 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.330 [2024-10-01 20:27:56.581610] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:01.330 [2024-10-01 20:27:56.581738] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:01.330 [2024-10-01 20:27:56.581817] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:01.589 [2024-10-01 20:27:56.584391] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:01.589 "name": "Existed_Raid", 00:29:01.589 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:01.589 "strip_size_kb": 64, 00:29:01.589 "state": "configuring", 00:29:01.589 "raid_level": "raid5f", 00:29:01.589 "superblock": false, 00:29:01.589 "num_base_bdevs": 3, 00:29:01.589 "num_base_bdevs_discovered": 2, 00:29:01.589 "num_base_bdevs_operational": 3, 00:29:01.589 "base_bdevs_list": [ 00:29:01.589 { 00:29:01.589 "name": "BaseBdev1", 00:29:01.589 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:01.589 "is_configured": false, 00:29:01.589 "data_offset": 0, 00:29:01.589 "data_size": 0 00:29:01.589 }, 00:29:01.589 { 00:29:01.589 "name": "BaseBdev2", 00:29:01.589 "uuid": "dc746513-7297-412c-8857-dbad6fd7e130", 00:29:01.589 "is_configured": true, 00:29:01.589 "data_offset": 0, 00:29:01.589 "data_size": 65536 00:29:01.589 }, 00:29:01.589 { 00:29:01.589 "name": "BaseBdev3", 00:29:01.589 "uuid": "0a6ba6d1-0664-49ed-a743-f3e9e704f553", 00:29:01.589 "is_configured": true, 00:29:01.589 "data_offset": 0, 00:29:01.589 "data_size": 65536 00:29:01.589 } 00:29:01.589 ] 00:29:01.589 }' 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:01.589 20:27:56 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.847 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:29:01.847 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:01.847 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:01.847 [2024-10-01 20:27:57.093798] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:02.106 "name": "Existed_Raid", 00:29:02.106 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:02.106 "strip_size_kb": 64, 00:29:02.106 "state": "configuring", 00:29:02.106 "raid_level": "raid5f", 00:29:02.106 "superblock": false, 00:29:02.106 "num_base_bdevs": 3, 00:29:02.106 "num_base_bdevs_discovered": 1, 00:29:02.106 "num_base_bdevs_operational": 3, 00:29:02.106 "base_bdevs_list": [ 00:29:02.106 { 00:29:02.106 "name": "BaseBdev1", 00:29:02.106 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:02.106 "is_configured": false, 00:29:02.106 "data_offset": 0, 00:29:02.106 "data_size": 0 00:29:02.106 }, 00:29:02.106 { 00:29:02.106 "name": null, 00:29:02.106 "uuid": "dc746513-7297-412c-8857-dbad6fd7e130", 00:29:02.106 "is_configured": false, 00:29:02.106 "data_offset": 0, 00:29:02.106 "data_size": 65536 00:29:02.106 }, 00:29:02.106 { 00:29:02.106 "name": "BaseBdev3", 00:29:02.106 "uuid": "0a6ba6d1-0664-49ed-a743-f3e9e704f553", 00:29:02.106 "is_configured": true, 00:29:02.106 "data_offset": 0, 00:29:02.106 "data_size": 65536 00:29:02.106 } 00:29:02.106 ] 00:29:02.106 }' 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:02.106 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:02.365 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:29:02.365 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:02.365 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:02.365 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:02.624 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:02.624 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:02.625 [2024-10-01 20:27:57.711731] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:02.625 BaseBdev1 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:02.625 [ 00:29:02.625 { 00:29:02.625 "name": "BaseBdev1", 00:29:02.625 "aliases": [ 00:29:02.625 "7a449454-fdd2-4bd8-814c-a09389b23f31" 00:29:02.625 ], 00:29:02.625 "product_name": "Malloc disk", 00:29:02.625 "block_size": 512, 00:29:02.625 "num_blocks": 65536, 00:29:02.625 "uuid": "7a449454-fdd2-4bd8-814c-a09389b23f31", 00:29:02.625 "assigned_rate_limits": { 00:29:02.625 "rw_ios_per_sec": 0, 00:29:02.625 "rw_mbytes_per_sec": 0, 00:29:02.625 "r_mbytes_per_sec": 0, 00:29:02.625 "w_mbytes_per_sec": 0 00:29:02.625 }, 00:29:02.625 "claimed": true, 00:29:02.625 "claim_type": "exclusive_write", 00:29:02.625 "zoned": false, 00:29:02.625 "supported_io_types": { 00:29:02.625 "read": true, 00:29:02.625 "write": true, 00:29:02.625 "unmap": true, 00:29:02.625 "flush": true, 00:29:02.625 "reset": true, 00:29:02.625 "nvme_admin": false, 00:29:02.625 "nvme_io": false, 00:29:02.625 "nvme_io_md": false, 00:29:02.625 "write_zeroes": true, 00:29:02.625 "zcopy": true, 00:29:02.625 "get_zone_info": false, 00:29:02.625 "zone_management": false, 00:29:02.625 "zone_append": false, 00:29:02.625 "compare": false, 00:29:02.625 "compare_and_write": false, 00:29:02.625 "abort": true, 00:29:02.625 "seek_hole": false, 00:29:02.625 "seek_data": false, 00:29:02.625 "copy": true, 00:29:02.625 "nvme_iov_md": false 00:29:02.625 }, 00:29:02.625 "memory_domains": [ 00:29:02.625 { 00:29:02.625 "dma_device_id": "system", 00:29:02.625 "dma_device_type": 1 00:29:02.625 }, 00:29:02.625 { 00:29:02.625 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:02.625 "dma_device_type": 2 00:29:02.625 } 00:29:02.625 ], 00:29:02.625 "driver_specific": {} 00:29:02.625 } 00:29:02.625 ] 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:02.625 "name": "Existed_Raid", 00:29:02.625 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:02.625 "strip_size_kb": 64, 00:29:02.625 "state": "configuring", 00:29:02.625 "raid_level": "raid5f", 00:29:02.625 "superblock": false, 00:29:02.625 "num_base_bdevs": 3, 00:29:02.625 "num_base_bdevs_discovered": 2, 00:29:02.625 "num_base_bdevs_operational": 3, 00:29:02.625 "base_bdevs_list": [ 00:29:02.625 { 00:29:02.625 "name": "BaseBdev1", 00:29:02.625 "uuid": "7a449454-fdd2-4bd8-814c-a09389b23f31", 00:29:02.625 "is_configured": true, 00:29:02.625 "data_offset": 0, 00:29:02.625 "data_size": 65536 00:29:02.625 }, 00:29:02.625 { 00:29:02.625 "name": null, 00:29:02.625 "uuid": "dc746513-7297-412c-8857-dbad6fd7e130", 00:29:02.625 "is_configured": false, 00:29:02.625 "data_offset": 0, 00:29:02.625 "data_size": 65536 00:29:02.625 }, 00:29:02.625 { 00:29:02.625 "name": "BaseBdev3", 00:29:02.625 "uuid": "0a6ba6d1-0664-49ed-a743-f3e9e704f553", 00:29:02.625 "is_configured": true, 00:29:02.625 "data_offset": 0, 00:29:02.625 "data_size": 65536 00:29:02.625 } 00:29:02.625 ] 00:29:02.625 }' 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:02.625 20:27:57 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:03.194 [2024-10-01 20:27:58.320060] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:03.194 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:03.195 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:03.195 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:03.195 "name": "Existed_Raid", 00:29:03.195 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:03.195 "strip_size_kb": 64, 00:29:03.195 "state": "configuring", 00:29:03.195 "raid_level": "raid5f", 00:29:03.195 "superblock": false, 00:29:03.195 "num_base_bdevs": 3, 00:29:03.195 "num_base_bdevs_discovered": 1, 00:29:03.195 "num_base_bdevs_operational": 3, 00:29:03.195 "base_bdevs_list": [ 00:29:03.195 { 00:29:03.195 "name": "BaseBdev1", 00:29:03.195 "uuid": "7a449454-fdd2-4bd8-814c-a09389b23f31", 00:29:03.195 "is_configured": true, 00:29:03.195 "data_offset": 0, 00:29:03.195 "data_size": 65536 00:29:03.195 }, 00:29:03.195 { 00:29:03.195 "name": null, 00:29:03.195 "uuid": "dc746513-7297-412c-8857-dbad6fd7e130", 00:29:03.195 "is_configured": false, 00:29:03.195 "data_offset": 0, 00:29:03.195 "data_size": 65536 00:29:03.195 }, 00:29:03.195 { 00:29:03.195 "name": null, 00:29:03.195 "uuid": "0a6ba6d1-0664-49ed-a743-f3e9e704f553", 00:29:03.195 "is_configured": false, 00:29:03.195 "data_offset": 0, 00:29:03.195 "data_size": 65536 00:29:03.195 } 00:29:03.195 ] 00:29:03.195 }' 00:29:03.195 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:03.195 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:03.763 [2024-10-01 20:27:58.916257] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:03.763 "name": "Existed_Raid", 00:29:03.763 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:03.763 "strip_size_kb": 64, 00:29:03.763 "state": "configuring", 00:29:03.763 "raid_level": "raid5f", 00:29:03.763 "superblock": false, 00:29:03.763 "num_base_bdevs": 3, 00:29:03.763 "num_base_bdevs_discovered": 2, 00:29:03.763 "num_base_bdevs_operational": 3, 00:29:03.763 "base_bdevs_list": [ 00:29:03.763 { 00:29:03.763 "name": "BaseBdev1", 00:29:03.763 "uuid": "7a449454-fdd2-4bd8-814c-a09389b23f31", 00:29:03.763 "is_configured": true, 00:29:03.763 "data_offset": 0, 00:29:03.763 "data_size": 65536 00:29:03.763 }, 00:29:03.763 { 00:29:03.763 "name": null, 00:29:03.763 "uuid": "dc746513-7297-412c-8857-dbad6fd7e130", 00:29:03.763 "is_configured": false, 00:29:03.763 "data_offset": 0, 00:29:03.763 "data_size": 65536 00:29:03.763 }, 00:29:03.763 { 00:29:03.763 "name": "BaseBdev3", 00:29:03.763 "uuid": "0a6ba6d1-0664-49ed-a743-f3e9e704f553", 00:29:03.763 "is_configured": true, 00:29:03.763 "data_offset": 0, 00:29:03.763 "data_size": 65536 00:29:03.763 } 00:29:03.763 ] 00:29:03.763 }' 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:03.763 20:27:58 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:04.330 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:29:04.330 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:04.330 20:27:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:04.330 20:27:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:04.330 20:27:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:04.330 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:29:04.330 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:29:04.330 20:27:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:04.330 20:27:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:04.330 [2024-10-01 20:27:59.492516] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:04.590 "name": "Existed_Raid", 00:29:04.590 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:04.590 "strip_size_kb": 64, 00:29:04.590 "state": "configuring", 00:29:04.590 "raid_level": "raid5f", 00:29:04.590 "superblock": false, 00:29:04.590 "num_base_bdevs": 3, 00:29:04.590 "num_base_bdevs_discovered": 1, 00:29:04.590 "num_base_bdevs_operational": 3, 00:29:04.590 "base_bdevs_list": [ 00:29:04.590 { 00:29:04.590 "name": null, 00:29:04.590 "uuid": "7a449454-fdd2-4bd8-814c-a09389b23f31", 00:29:04.590 "is_configured": false, 00:29:04.590 "data_offset": 0, 00:29:04.590 "data_size": 65536 00:29:04.590 }, 00:29:04.590 { 00:29:04.590 "name": null, 00:29:04.590 "uuid": "dc746513-7297-412c-8857-dbad6fd7e130", 00:29:04.590 "is_configured": false, 00:29:04.590 "data_offset": 0, 00:29:04.590 "data_size": 65536 00:29:04.590 }, 00:29:04.590 { 00:29:04.590 "name": "BaseBdev3", 00:29:04.590 "uuid": "0a6ba6d1-0664-49ed-a743-f3e9e704f553", 00:29:04.590 "is_configured": true, 00:29:04.590 "data_offset": 0, 00:29:04.590 "data_size": 65536 00:29:04.590 } 00:29:04.590 ] 00:29:04.590 }' 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:04.590 20:27:59 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:05.159 [2024-10-01 20:28:00.187474] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:05.159 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:05.160 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:05.160 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:05.160 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:05.160 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:05.160 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:05.160 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:05.160 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:05.160 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:05.160 "name": "Existed_Raid", 00:29:05.160 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:05.160 "strip_size_kb": 64, 00:29:05.160 "state": "configuring", 00:29:05.160 "raid_level": "raid5f", 00:29:05.160 "superblock": false, 00:29:05.160 "num_base_bdevs": 3, 00:29:05.160 "num_base_bdevs_discovered": 2, 00:29:05.160 "num_base_bdevs_operational": 3, 00:29:05.160 "base_bdevs_list": [ 00:29:05.160 { 00:29:05.160 "name": null, 00:29:05.160 "uuid": "7a449454-fdd2-4bd8-814c-a09389b23f31", 00:29:05.160 "is_configured": false, 00:29:05.160 "data_offset": 0, 00:29:05.160 "data_size": 65536 00:29:05.160 }, 00:29:05.160 { 00:29:05.160 "name": "BaseBdev2", 00:29:05.160 "uuid": "dc746513-7297-412c-8857-dbad6fd7e130", 00:29:05.160 "is_configured": true, 00:29:05.160 "data_offset": 0, 00:29:05.160 "data_size": 65536 00:29:05.160 }, 00:29:05.160 { 00:29:05.160 "name": "BaseBdev3", 00:29:05.160 "uuid": "0a6ba6d1-0664-49ed-a743-f3e9e704f553", 00:29:05.160 "is_configured": true, 00:29:05.160 "data_offset": 0, 00:29:05.160 "data_size": 65536 00:29:05.160 } 00:29:05.160 ] 00:29:05.160 }' 00:29:05.160 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:05.160 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 7a449454-fdd2-4bd8-814c-a09389b23f31 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:05.732 [2024-10-01 20:28:00.861336] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:29:05.732 [2024-10-01 20:28:00.861580] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:29:05.732 [2024-10-01 20:28:00.861611] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:29:05.732 [2024-10-01 20:28:00.862024] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:29:05.732 [2024-10-01 20:28:00.868466] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:29:05.732 [2024-10-01 20:28:00.868640] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:29:05.732 [2024-10-01 20:28:00.869122] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:05.732 NewBaseBdev 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:05.732 [ 00:29:05.732 { 00:29:05.732 "name": "NewBaseBdev", 00:29:05.732 "aliases": [ 00:29:05.732 "7a449454-fdd2-4bd8-814c-a09389b23f31" 00:29:05.732 ], 00:29:05.732 "product_name": "Malloc disk", 00:29:05.732 "block_size": 512, 00:29:05.732 "num_blocks": 65536, 00:29:05.732 "uuid": "7a449454-fdd2-4bd8-814c-a09389b23f31", 00:29:05.732 "assigned_rate_limits": { 00:29:05.732 "rw_ios_per_sec": 0, 00:29:05.732 "rw_mbytes_per_sec": 0, 00:29:05.732 "r_mbytes_per_sec": 0, 00:29:05.732 "w_mbytes_per_sec": 0 00:29:05.732 }, 00:29:05.732 "claimed": true, 00:29:05.732 "claim_type": "exclusive_write", 00:29:05.732 "zoned": false, 00:29:05.732 "supported_io_types": { 00:29:05.732 "read": true, 00:29:05.732 "write": true, 00:29:05.732 "unmap": true, 00:29:05.732 "flush": true, 00:29:05.732 "reset": true, 00:29:05.732 "nvme_admin": false, 00:29:05.732 "nvme_io": false, 00:29:05.732 "nvme_io_md": false, 00:29:05.732 "write_zeroes": true, 00:29:05.732 "zcopy": true, 00:29:05.732 "get_zone_info": false, 00:29:05.732 "zone_management": false, 00:29:05.732 "zone_append": false, 00:29:05.732 "compare": false, 00:29:05.732 "compare_and_write": false, 00:29:05.732 "abort": true, 00:29:05.732 "seek_hole": false, 00:29:05.732 "seek_data": false, 00:29:05.732 "copy": true, 00:29:05.732 "nvme_iov_md": false 00:29:05.732 }, 00:29:05.732 "memory_domains": [ 00:29:05.732 { 00:29:05.732 "dma_device_id": "system", 00:29:05.732 "dma_device_type": 1 00:29:05.732 }, 00:29:05.732 { 00:29:05.732 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:05.732 "dma_device_type": 2 00:29:05.732 } 00:29:05.732 ], 00:29:05.732 "driver_specific": {} 00:29:05.732 } 00:29:05.732 ] 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:05.732 "name": "Existed_Raid", 00:29:05.732 "uuid": "861d7e1f-ec22-43e1-93c2-91f948bf56b3", 00:29:05.732 "strip_size_kb": 64, 00:29:05.732 "state": "online", 00:29:05.732 "raid_level": "raid5f", 00:29:05.732 "superblock": false, 00:29:05.732 "num_base_bdevs": 3, 00:29:05.732 "num_base_bdevs_discovered": 3, 00:29:05.732 "num_base_bdevs_operational": 3, 00:29:05.732 "base_bdevs_list": [ 00:29:05.732 { 00:29:05.732 "name": "NewBaseBdev", 00:29:05.732 "uuid": "7a449454-fdd2-4bd8-814c-a09389b23f31", 00:29:05.732 "is_configured": true, 00:29:05.732 "data_offset": 0, 00:29:05.732 "data_size": 65536 00:29:05.732 }, 00:29:05.732 { 00:29:05.732 "name": "BaseBdev2", 00:29:05.732 "uuid": "dc746513-7297-412c-8857-dbad6fd7e130", 00:29:05.732 "is_configured": true, 00:29:05.732 "data_offset": 0, 00:29:05.732 "data_size": 65536 00:29:05.732 }, 00:29:05.732 { 00:29:05.732 "name": "BaseBdev3", 00:29:05.732 "uuid": "0a6ba6d1-0664-49ed-a743-f3e9e704f553", 00:29:05.732 "is_configured": true, 00:29:05.732 "data_offset": 0, 00:29:05.732 "data_size": 65536 00:29:05.732 } 00:29:05.732 ] 00:29:05.732 }' 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:05.732 20:28:00 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:06.300 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:29:06.301 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:29:06.301 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:06.301 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:06.301 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:29:06.301 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:06.301 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:29:06.301 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:06.301 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:06.301 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:06.301 [2024-10-01 20:28:01.443653] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:06.301 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:06.301 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:06.301 "name": "Existed_Raid", 00:29:06.301 "aliases": [ 00:29:06.301 "861d7e1f-ec22-43e1-93c2-91f948bf56b3" 00:29:06.301 ], 00:29:06.301 "product_name": "Raid Volume", 00:29:06.301 "block_size": 512, 00:29:06.301 "num_blocks": 131072, 00:29:06.301 "uuid": "861d7e1f-ec22-43e1-93c2-91f948bf56b3", 00:29:06.301 "assigned_rate_limits": { 00:29:06.301 "rw_ios_per_sec": 0, 00:29:06.301 "rw_mbytes_per_sec": 0, 00:29:06.301 "r_mbytes_per_sec": 0, 00:29:06.301 "w_mbytes_per_sec": 0 00:29:06.301 }, 00:29:06.301 "claimed": false, 00:29:06.301 "zoned": false, 00:29:06.301 "supported_io_types": { 00:29:06.301 "read": true, 00:29:06.301 "write": true, 00:29:06.301 "unmap": false, 00:29:06.301 "flush": false, 00:29:06.301 "reset": true, 00:29:06.301 "nvme_admin": false, 00:29:06.301 "nvme_io": false, 00:29:06.301 "nvme_io_md": false, 00:29:06.301 "write_zeroes": true, 00:29:06.301 "zcopy": false, 00:29:06.301 "get_zone_info": false, 00:29:06.301 "zone_management": false, 00:29:06.301 "zone_append": false, 00:29:06.301 "compare": false, 00:29:06.301 "compare_and_write": false, 00:29:06.301 "abort": false, 00:29:06.301 "seek_hole": false, 00:29:06.301 "seek_data": false, 00:29:06.301 "copy": false, 00:29:06.301 "nvme_iov_md": false 00:29:06.301 }, 00:29:06.301 "driver_specific": { 00:29:06.301 "raid": { 00:29:06.301 "uuid": "861d7e1f-ec22-43e1-93c2-91f948bf56b3", 00:29:06.301 "strip_size_kb": 64, 00:29:06.301 "state": "online", 00:29:06.301 "raid_level": "raid5f", 00:29:06.301 "superblock": false, 00:29:06.301 "num_base_bdevs": 3, 00:29:06.301 "num_base_bdevs_discovered": 3, 00:29:06.301 "num_base_bdevs_operational": 3, 00:29:06.301 "base_bdevs_list": [ 00:29:06.301 { 00:29:06.301 "name": "NewBaseBdev", 00:29:06.301 "uuid": "7a449454-fdd2-4bd8-814c-a09389b23f31", 00:29:06.301 "is_configured": true, 00:29:06.301 "data_offset": 0, 00:29:06.301 "data_size": 65536 00:29:06.301 }, 00:29:06.301 { 00:29:06.301 "name": "BaseBdev2", 00:29:06.301 "uuid": "dc746513-7297-412c-8857-dbad6fd7e130", 00:29:06.301 "is_configured": true, 00:29:06.301 "data_offset": 0, 00:29:06.301 "data_size": 65536 00:29:06.301 }, 00:29:06.301 { 00:29:06.301 "name": "BaseBdev3", 00:29:06.301 "uuid": "0a6ba6d1-0664-49ed-a743-f3e9e704f553", 00:29:06.301 "is_configured": true, 00:29:06.301 "data_offset": 0, 00:29:06.301 "data_size": 65536 00:29:06.301 } 00:29:06.301 ] 00:29:06.301 } 00:29:06.301 } 00:29:06.301 }' 00:29:06.301 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:06.301 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:29:06.301 BaseBdev2 00:29:06.301 BaseBdev3' 00:29:06.301 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:06.560 [2024-10-01 20:28:01.759473] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:06.560 [2024-10-01 20:28:01.759638] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:06.560 [2024-10-01 20:28:01.759890] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:06.560 [2024-10-01 20:28:01.760418] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:06.560 [2024-10-01 20:28:01.760554] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 81088 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 81088 ']' 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # kill -0 81088 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@955 -- # uname 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 81088 00:29:06.560 killing process with pid 81088 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 81088' 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@969 -- # kill 81088 00:29:06.560 [2024-10-01 20:28:01.802501] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:29:06.560 20:28:01 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@974 -- # wait 81088 00:29:07.126 [2024-10-01 20:28:02.100628] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:29:09.029 20:28:04 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:29:09.029 00:29:09.029 real 0m13.266s 00:29:09.029 user 0m20.885s 00:29:09.029 sys 0m1.982s 00:29:09.029 20:28:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:29:09.029 ************************************ 00:29:09.029 END TEST raid5f_state_function_test 00:29:09.029 ************************************ 00:29:09.029 20:28:04 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:29:09.029 20:28:04 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 3 true 00:29:09.029 20:28:04 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:29:09.029 20:28:04 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:29:09.029 20:28:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:29:09.029 ************************************ 00:29:09.029 START TEST raid5f_state_function_test_sb 00:29:09.029 ************************************ 00:29:09.029 20:28:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid5f 3 true 00:29:09.029 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:29:09.029 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:29:09.029 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:29:09.029 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:29:09.029 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:29:09.029 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:09.029 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:29:09.029 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:29:09.030 Process raid pid: 81732 00:29:09.030 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=81732 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 81732' 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 81732 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 81732 ']' 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:29:09.030 20:28:04 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:09.288 [2024-10-01 20:28:04.365112] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:29:09.288 [2024-10-01 20:28:04.365635] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:29:09.547 [2024-10-01 20:28:04.549341] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:09.806 [2024-10-01 20:28:04.847512] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:29:10.065 [2024-10-01 20:28:05.082218] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:10.065 [2024-10-01 20:28:05.082428] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:10.325 [2024-10-01 20:28:05.550143] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:10.325 [2024-10-01 20:28:05.550354] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:10.325 [2024-10-01 20:28:05.550492] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:10.325 [2024-10-01 20:28:05.550557] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:10.325 [2024-10-01 20:28:05.550674] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:10.325 [2024-10-01 20:28:05.550754] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:10.325 20:28:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:10.584 20:28:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:10.584 "name": "Existed_Raid", 00:29:10.584 "uuid": "e8fc877c-0498-41e3-8c99-2b105b56ce4a", 00:29:10.584 "strip_size_kb": 64, 00:29:10.584 "state": "configuring", 00:29:10.584 "raid_level": "raid5f", 00:29:10.584 "superblock": true, 00:29:10.584 "num_base_bdevs": 3, 00:29:10.584 "num_base_bdevs_discovered": 0, 00:29:10.584 "num_base_bdevs_operational": 3, 00:29:10.584 "base_bdevs_list": [ 00:29:10.584 { 00:29:10.584 "name": "BaseBdev1", 00:29:10.584 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:10.584 "is_configured": false, 00:29:10.584 "data_offset": 0, 00:29:10.584 "data_size": 0 00:29:10.584 }, 00:29:10.584 { 00:29:10.584 "name": "BaseBdev2", 00:29:10.584 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:10.584 "is_configured": false, 00:29:10.584 "data_offset": 0, 00:29:10.584 "data_size": 0 00:29:10.584 }, 00:29:10.584 { 00:29:10.584 "name": "BaseBdev3", 00:29:10.584 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:10.584 "is_configured": false, 00:29:10.584 "data_offset": 0, 00:29:10.584 "data_size": 0 00:29:10.584 } 00:29:10.584 ] 00:29:10.584 }' 00:29:10.584 20:28:05 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:10.584 20:28:05 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:11.153 [2024-10-01 20:28:06.118191] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:11.153 [2024-10-01 20:28:06.118242] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:11.153 [2024-10-01 20:28:06.126220] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:11.153 [2024-10-01 20:28:06.126387] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:11.153 [2024-10-01 20:28:06.126508] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:11.153 [2024-10-01 20:28:06.126572] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:11.153 [2024-10-01 20:28:06.126711] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:11.153 [2024-10-01 20:28:06.126789] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:11.153 [2024-10-01 20:28:06.172786] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:11.153 BaseBdev1 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:11.153 [ 00:29:11.153 { 00:29:11.153 "name": "BaseBdev1", 00:29:11.153 "aliases": [ 00:29:11.153 "f596e0e1-2d89-4a3d-8e22-efdcdbc47bab" 00:29:11.153 ], 00:29:11.153 "product_name": "Malloc disk", 00:29:11.153 "block_size": 512, 00:29:11.153 "num_blocks": 65536, 00:29:11.153 "uuid": "f596e0e1-2d89-4a3d-8e22-efdcdbc47bab", 00:29:11.153 "assigned_rate_limits": { 00:29:11.153 "rw_ios_per_sec": 0, 00:29:11.153 "rw_mbytes_per_sec": 0, 00:29:11.153 "r_mbytes_per_sec": 0, 00:29:11.153 "w_mbytes_per_sec": 0 00:29:11.153 }, 00:29:11.153 "claimed": true, 00:29:11.153 "claim_type": "exclusive_write", 00:29:11.153 "zoned": false, 00:29:11.153 "supported_io_types": { 00:29:11.153 "read": true, 00:29:11.153 "write": true, 00:29:11.153 "unmap": true, 00:29:11.153 "flush": true, 00:29:11.153 "reset": true, 00:29:11.153 "nvme_admin": false, 00:29:11.153 "nvme_io": false, 00:29:11.153 "nvme_io_md": false, 00:29:11.153 "write_zeroes": true, 00:29:11.153 "zcopy": true, 00:29:11.153 "get_zone_info": false, 00:29:11.153 "zone_management": false, 00:29:11.153 "zone_append": false, 00:29:11.153 "compare": false, 00:29:11.153 "compare_and_write": false, 00:29:11.153 "abort": true, 00:29:11.153 "seek_hole": false, 00:29:11.153 "seek_data": false, 00:29:11.153 "copy": true, 00:29:11.153 "nvme_iov_md": false 00:29:11.153 }, 00:29:11.153 "memory_domains": [ 00:29:11.153 { 00:29:11.153 "dma_device_id": "system", 00:29:11.153 "dma_device_type": 1 00:29:11.153 }, 00:29:11.153 { 00:29:11.153 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:11.153 "dma_device_type": 2 00:29:11.153 } 00:29:11.153 ], 00:29:11.153 "driver_specific": {} 00:29:11.153 } 00:29:11.153 ] 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:11.153 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:11.154 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:11.154 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:11.154 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:11.154 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:11.154 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:11.154 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:11.154 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.154 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:11.154 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:11.154 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.154 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:11.154 "name": "Existed_Raid", 00:29:11.154 "uuid": "4cb857e2-8570-497f-90ec-398c60382732", 00:29:11.154 "strip_size_kb": 64, 00:29:11.154 "state": "configuring", 00:29:11.154 "raid_level": "raid5f", 00:29:11.154 "superblock": true, 00:29:11.154 "num_base_bdevs": 3, 00:29:11.154 "num_base_bdevs_discovered": 1, 00:29:11.154 "num_base_bdevs_operational": 3, 00:29:11.154 "base_bdevs_list": [ 00:29:11.154 { 00:29:11.154 "name": "BaseBdev1", 00:29:11.154 "uuid": "f596e0e1-2d89-4a3d-8e22-efdcdbc47bab", 00:29:11.154 "is_configured": true, 00:29:11.154 "data_offset": 2048, 00:29:11.154 "data_size": 63488 00:29:11.154 }, 00:29:11.154 { 00:29:11.154 "name": "BaseBdev2", 00:29:11.154 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:11.154 "is_configured": false, 00:29:11.154 "data_offset": 0, 00:29:11.154 "data_size": 0 00:29:11.154 }, 00:29:11.154 { 00:29:11.154 "name": "BaseBdev3", 00:29:11.154 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:11.154 "is_configured": false, 00:29:11.154 "data_offset": 0, 00:29:11.154 "data_size": 0 00:29:11.154 } 00:29:11.154 ] 00:29:11.154 }' 00:29:11.154 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:11.154 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:11.721 [2024-10-01 20:28:06.709016] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:11.721 [2024-10-01 20:28:06.709255] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:11.721 [2024-10-01 20:28:06.717051] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:11.721 [2024-10-01 20:28:06.719850] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:29:11.721 [2024-10-01 20:28:06.720027] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:29:11.721 [2024-10-01 20:28:06.720154] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:29:11.721 [2024-10-01 20:28:06.720218] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:11.721 "name": "Existed_Raid", 00:29:11.721 "uuid": "adf17e27-4d3a-4d0f-8b13-ee3e1e53618b", 00:29:11.721 "strip_size_kb": 64, 00:29:11.721 "state": "configuring", 00:29:11.721 "raid_level": "raid5f", 00:29:11.721 "superblock": true, 00:29:11.721 "num_base_bdevs": 3, 00:29:11.721 "num_base_bdevs_discovered": 1, 00:29:11.721 "num_base_bdevs_operational": 3, 00:29:11.721 "base_bdevs_list": [ 00:29:11.721 { 00:29:11.721 "name": "BaseBdev1", 00:29:11.721 "uuid": "f596e0e1-2d89-4a3d-8e22-efdcdbc47bab", 00:29:11.721 "is_configured": true, 00:29:11.721 "data_offset": 2048, 00:29:11.721 "data_size": 63488 00:29:11.721 }, 00:29:11.721 { 00:29:11.721 "name": "BaseBdev2", 00:29:11.721 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:11.721 "is_configured": false, 00:29:11.721 "data_offset": 0, 00:29:11.721 "data_size": 0 00:29:11.721 }, 00:29:11.721 { 00:29:11.721 "name": "BaseBdev3", 00:29:11.721 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:11.721 "is_configured": false, 00:29:11.721 "data_offset": 0, 00:29:11.721 "data_size": 0 00:29:11.721 } 00:29:11.721 ] 00:29:11.721 }' 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:11.721 20:28:06 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:12.289 [2024-10-01 20:28:07.279470] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:12.289 BaseBdev2 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:12.289 [ 00:29:12.289 { 00:29:12.289 "name": "BaseBdev2", 00:29:12.289 "aliases": [ 00:29:12.289 "17da0796-5aad-4891-86fd-6b6d5098118e" 00:29:12.289 ], 00:29:12.289 "product_name": "Malloc disk", 00:29:12.289 "block_size": 512, 00:29:12.289 "num_blocks": 65536, 00:29:12.289 "uuid": "17da0796-5aad-4891-86fd-6b6d5098118e", 00:29:12.289 "assigned_rate_limits": { 00:29:12.289 "rw_ios_per_sec": 0, 00:29:12.289 "rw_mbytes_per_sec": 0, 00:29:12.289 "r_mbytes_per_sec": 0, 00:29:12.289 "w_mbytes_per_sec": 0 00:29:12.289 }, 00:29:12.289 "claimed": true, 00:29:12.289 "claim_type": "exclusive_write", 00:29:12.289 "zoned": false, 00:29:12.289 "supported_io_types": { 00:29:12.289 "read": true, 00:29:12.289 "write": true, 00:29:12.289 "unmap": true, 00:29:12.289 "flush": true, 00:29:12.289 "reset": true, 00:29:12.289 "nvme_admin": false, 00:29:12.289 "nvme_io": false, 00:29:12.289 "nvme_io_md": false, 00:29:12.289 "write_zeroes": true, 00:29:12.289 "zcopy": true, 00:29:12.289 "get_zone_info": false, 00:29:12.289 "zone_management": false, 00:29:12.289 "zone_append": false, 00:29:12.289 "compare": false, 00:29:12.289 "compare_and_write": false, 00:29:12.289 "abort": true, 00:29:12.289 "seek_hole": false, 00:29:12.289 "seek_data": false, 00:29:12.289 "copy": true, 00:29:12.289 "nvme_iov_md": false 00:29:12.289 }, 00:29:12.289 "memory_domains": [ 00:29:12.289 { 00:29:12.289 "dma_device_id": "system", 00:29:12.289 "dma_device_type": 1 00:29:12.289 }, 00:29:12.289 { 00:29:12.289 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:12.289 "dma_device_type": 2 00:29:12.289 } 00:29:12.289 ], 00:29:12.289 "driver_specific": {} 00:29:12.289 } 00:29:12.289 ] 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:12.289 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:12.289 "name": "Existed_Raid", 00:29:12.289 "uuid": "adf17e27-4d3a-4d0f-8b13-ee3e1e53618b", 00:29:12.289 "strip_size_kb": 64, 00:29:12.289 "state": "configuring", 00:29:12.289 "raid_level": "raid5f", 00:29:12.289 "superblock": true, 00:29:12.289 "num_base_bdevs": 3, 00:29:12.289 "num_base_bdevs_discovered": 2, 00:29:12.289 "num_base_bdevs_operational": 3, 00:29:12.289 "base_bdevs_list": [ 00:29:12.289 { 00:29:12.289 "name": "BaseBdev1", 00:29:12.289 "uuid": "f596e0e1-2d89-4a3d-8e22-efdcdbc47bab", 00:29:12.289 "is_configured": true, 00:29:12.289 "data_offset": 2048, 00:29:12.289 "data_size": 63488 00:29:12.289 }, 00:29:12.289 { 00:29:12.290 "name": "BaseBdev2", 00:29:12.290 "uuid": "17da0796-5aad-4891-86fd-6b6d5098118e", 00:29:12.290 "is_configured": true, 00:29:12.290 "data_offset": 2048, 00:29:12.290 "data_size": 63488 00:29:12.290 }, 00:29:12.290 { 00:29:12.290 "name": "BaseBdev3", 00:29:12.290 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:12.290 "is_configured": false, 00:29:12.290 "data_offset": 0, 00:29:12.290 "data_size": 0 00:29:12.290 } 00:29:12.290 ] 00:29:12.290 }' 00:29:12.290 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:12.290 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:12.858 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:29:12.858 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:12.858 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:12.858 [2024-10-01 20:28:07.878263] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:12.858 [2024-10-01 20:28:07.878610] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:29:12.858 [2024-10-01 20:28:07.878645] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:29:12.858 BaseBdev3 00:29:12.858 [2024-10-01 20:28:07.879183] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:29:12.858 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:12.858 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:29:12.858 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:29:12.858 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:29:12.858 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:29:12.858 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:29:12.858 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:29:12.858 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:29:12.858 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:12.858 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:12.859 [2024-10-01 20:28:07.884646] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:29:12.859 [2024-10-01 20:28:07.884844] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:29:12.859 [2024-10-01 20:28:07.885373] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:12.859 [ 00:29:12.859 { 00:29:12.859 "name": "BaseBdev3", 00:29:12.859 "aliases": [ 00:29:12.859 "80a6bf08-fe14-496f-8580-85f4c78f8072" 00:29:12.859 ], 00:29:12.859 "product_name": "Malloc disk", 00:29:12.859 "block_size": 512, 00:29:12.859 "num_blocks": 65536, 00:29:12.859 "uuid": "80a6bf08-fe14-496f-8580-85f4c78f8072", 00:29:12.859 "assigned_rate_limits": { 00:29:12.859 "rw_ios_per_sec": 0, 00:29:12.859 "rw_mbytes_per_sec": 0, 00:29:12.859 "r_mbytes_per_sec": 0, 00:29:12.859 "w_mbytes_per_sec": 0 00:29:12.859 }, 00:29:12.859 "claimed": true, 00:29:12.859 "claim_type": "exclusive_write", 00:29:12.859 "zoned": false, 00:29:12.859 "supported_io_types": { 00:29:12.859 "read": true, 00:29:12.859 "write": true, 00:29:12.859 "unmap": true, 00:29:12.859 "flush": true, 00:29:12.859 "reset": true, 00:29:12.859 "nvme_admin": false, 00:29:12.859 "nvme_io": false, 00:29:12.859 "nvme_io_md": false, 00:29:12.859 "write_zeroes": true, 00:29:12.859 "zcopy": true, 00:29:12.859 "get_zone_info": false, 00:29:12.859 "zone_management": false, 00:29:12.859 "zone_append": false, 00:29:12.859 "compare": false, 00:29:12.859 "compare_and_write": false, 00:29:12.859 "abort": true, 00:29:12.859 "seek_hole": false, 00:29:12.859 "seek_data": false, 00:29:12.859 "copy": true, 00:29:12.859 "nvme_iov_md": false 00:29:12.859 }, 00:29:12.859 "memory_domains": [ 00:29:12.859 { 00:29:12.859 "dma_device_id": "system", 00:29:12.859 "dma_device_type": 1 00:29:12.859 }, 00:29:12.859 { 00:29:12.859 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:12.859 "dma_device_type": 2 00:29:12.859 } 00:29:12.859 ], 00:29:12.859 "driver_specific": {} 00:29:12.859 } 00:29:12.859 ] 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:12.859 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:12.860 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:12.860 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:12.860 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:12.860 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:12.860 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:12.860 "name": "Existed_Raid", 00:29:12.860 "uuid": "adf17e27-4d3a-4d0f-8b13-ee3e1e53618b", 00:29:12.860 "strip_size_kb": 64, 00:29:12.860 "state": "online", 00:29:12.860 "raid_level": "raid5f", 00:29:12.860 "superblock": true, 00:29:12.860 "num_base_bdevs": 3, 00:29:12.860 "num_base_bdevs_discovered": 3, 00:29:12.860 "num_base_bdevs_operational": 3, 00:29:12.860 "base_bdevs_list": [ 00:29:12.860 { 00:29:12.860 "name": "BaseBdev1", 00:29:12.860 "uuid": "f596e0e1-2d89-4a3d-8e22-efdcdbc47bab", 00:29:12.860 "is_configured": true, 00:29:12.860 "data_offset": 2048, 00:29:12.860 "data_size": 63488 00:29:12.860 }, 00:29:12.860 { 00:29:12.860 "name": "BaseBdev2", 00:29:12.860 "uuid": "17da0796-5aad-4891-86fd-6b6d5098118e", 00:29:12.860 "is_configured": true, 00:29:12.860 "data_offset": 2048, 00:29:12.860 "data_size": 63488 00:29:12.860 }, 00:29:12.860 { 00:29:12.860 "name": "BaseBdev3", 00:29:12.860 "uuid": "80a6bf08-fe14-496f-8580-85f4c78f8072", 00:29:12.860 "is_configured": true, 00:29:12.860 "data_offset": 2048, 00:29:12.860 "data_size": 63488 00:29:12.860 } 00:29:12.860 ] 00:29:12.860 }' 00:29:12.860 20:28:07 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:12.860 20:28:07 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:13.427 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:29:13.427 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:29:13.427 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:13.427 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:13.427 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:13.428 [2024-10-01 20:28:08.431761] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:13.428 "name": "Existed_Raid", 00:29:13.428 "aliases": [ 00:29:13.428 "adf17e27-4d3a-4d0f-8b13-ee3e1e53618b" 00:29:13.428 ], 00:29:13.428 "product_name": "Raid Volume", 00:29:13.428 "block_size": 512, 00:29:13.428 "num_blocks": 126976, 00:29:13.428 "uuid": "adf17e27-4d3a-4d0f-8b13-ee3e1e53618b", 00:29:13.428 "assigned_rate_limits": { 00:29:13.428 "rw_ios_per_sec": 0, 00:29:13.428 "rw_mbytes_per_sec": 0, 00:29:13.428 "r_mbytes_per_sec": 0, 00:29:13.428 "w_mbytes_per_sec": 0 00:29:13.428 }, 00:29:13.428 "claimed": false, 00:29:13.428 "zoned": false, 00:29:13.428 "supported_io_types": { 00:29:13.428 "read": true, 00:29:13.428 "write": true, 00:29:13.428 "unmap": false, 00:29:13.428 "flush": false, 00:29:13.428 "reset": true, 00:29:13.428 "nvme_admin": false, 00:29:13.428 "nvme_io": false, 00:29:13.428 "nvme_io_md": false, 00:29:13.428 "write_zeroes": true, 00:29:13.428 "zcopy": false, 00:29:13.428 "get_zone_info": false, 00:29:13.428 "zone_management": false, 00:29:13.428 "zone_append": false, 00:29:13.428 "compare": false, 00:29:13.428 "compare_and_write": false, 00:29:13.428 "abort": false, 00:29:13.428 "seek_hole": false, 00:29:13.428 "seek_data": false, 00:29:13.428 "copy": false, 00:29:13.428 "nvme_iov_md": false 00:29:13.428 }, 00:29:13.428 "driver_specific": { 00:29:13.428 "raid": { 00:29:13.428 "uuid": "adf17e27-4d3a-4d0f-8b13-ee3e1e53618b", 00:29:13.428 "strip_size_kb": 64, 00:29:13.428 "state": "online", 00:29:13.428 "raid_level": "raid5f", 00:29:13.428 "superblock": true, 00:29:13.428 "num_base_bdevs": 3, 00:29:13.428 "num_base_bdevs_discovered": 3, 00:29:13.428 "num_base_bdevs_operational": 3, 00:29:13.428 "base_bdevs_list": [ 00:29:13.428 { 00:29:13.428 "name": "BaseBdev1", 00:29:13.428 "uuid": "f596e0e1-2d89-4a3d-8e22-efdcdbc47bab", 00:29:13.428 "is_configured": true, 00:29:13.428 "data_offset": 2048, 00:29:13.428 "data_size": 63488 00:29:13.428 }, 00:29:13.428 { 00:29:13.428 "name": "BaseBdev2", 00:29:13.428 "uuid": "17da0796-5aad-4891-86fd-6b6d5098118e", 00:29:13.428 "is_configured": true, 00:29:13.428 "data_offset": 2048, 00:29:13.428 "data_size": 63488 00:29:13.428 }, 00:29:13.428 { 00:29:13.428 "name": "BaseBdev3", 00:29:13.428 "uuid": "80a6bf08-fe14-496f-8580-85f4c78f8072", 00:29:13.428 "is_configured": true, 00:29:13.428 "data_offset": 2048, 00:29:13.428 "data_size": 63488 00:29:13.428 } 00:29:13.428 ] 00:29:13.428 } 00:29:13.428 } 00:29:13.428 }' 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:29:13.428 BaseBdev2 00:29:13.428 BaseBdev3' 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:13.428 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:13.687 [2024-10-01 20:28:08.755610] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:13.687 "name": "Existed_Raid", 00:29:13.687 "uuid": "adf17e27-4d3a-4d0f-8b13-ee3e1e53618b", 00:29:13.687 "strip_size_kb": 64, 00:29:13.687 "state": "online", 00:29:13.687 "raid_level": "raid5f", 00:29:13.687 "superblock": true, 00:29:13.687 "num_base_bdevs": 3, 00:29:13.687 "num_base_bdevs_discovered": 2, 00:29:13.687 "num_base_bdevs_operational": 2, 00:29:13.687 "base_bdevs_list": [ 00:29:13.687 { 00:29:13.687 "name": null, 00:29:13.687 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:13.687 "is_configured": false, 00:29:13.687 "data_offset": 0, 00:29:13.687 "data_size": 63488 00:29:13.687 }, 00:29:13.687 { 00:29:13.687 "name": "BaseBdev2", 00:29:13.687 "uuid": "17da0796-5aad-4891-86fd-6b6d5098118e", 00:29:13.687 "is_configured": true, 00:29:13.687 "data_offset": 2048, 00:29:13.687 "data_size": 63488 00:29:13.687 }, 00:29:13.687 { 00:29:13.687 "name": "BaseBdev3", 00:29:13.687 "uuid": "80a6bf08-fe14-496f-8580-85f4c78f8072", 00:29:13.687 "is_configured": true, 00:29:13.687 "data_offset": 2048, 00:29:13.687 "data_size": 63488 00:29:13.687 } 00:29:13.687 ] 00:29:13.687 }' 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:13.687 20:28:08 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:14.254 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:29:14.254 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:14.254 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:29:14.254 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:14.254 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:14.254 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:14.254 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:14.254 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:29:14.254 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:29:14.254 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:29:14.254 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:14.254 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:14.254 [2024-10-01 20:28:09.427246] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:29:14.254 [2024-10-01 20:28:09.427484] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:14.513 [2024-10-01 20:28:09.530768] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:14.513 [2024-10-01 20:28:09.594815] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:29:14.513 [2024-10-01 20:28:09.595029] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:14.513 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:14.772 BaseBdev2 00:29:14.772 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:14.772 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:29:14.772 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:29:14.772 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:29:14.772 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:29:14.772 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:29:14.772 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:14.773 [ 00:29:14.773 { 00:29:14.773 "name": "BaseBdev2", 00:29:14.773 "aliases": [ 00:29:14.773 "8184849d-05c0-447c-b4fe-01784bc12d3f" 00:29:14.773 ], 00:29:14.773 "product_name": "Malloc disk", 00:29:14.773 "block_size": 512, 00:29:14.773 "num_blocks": 65536, 00:29:14.773 "uuid": "8184849d-05c0-447c-b4fe-01784bc12d3f", 00:29:14.773 "assigned_rate_limits": { 00:29:14.773 "rw_ios_per_sec": 0, 00:29:14.773 "rw_mbytes_per_sec": 0, 00:29:14.773 "r_mbytes_per_sec": 0, 00:29:14.773 "w_mbytes_per_sec": 0 00:29:14.773 }, 00:29:14.773 "claimed": false, 00:29:14.773 "zoned": false, 00:29:14.773 "supported_io_types": { 00:29:14.773 "read": true, 00:29:14.773 "write": true, 00:29:14.773 "unmap": true, 00:29:14.773 "flush": true, 00:29:14.773 "reset": true, 00:29:14.773 "nvme_admin": false, 00:29:14.773 "nvme_io": false, 00:29:14.773 "nvme_io_md": false, 00:29:14.773 "write_zeroes": true, 00:29:14.773 "zcopy": true, 00:29:14.773 "get_zone_info": false, 00:29:14.773 "zone_management": false, 00:29:14.773 "zone_append": false, 00:29:14.773 "compare": false, 00:29:14.773 "compare_and_write": false, 00:29:14.773 "abort": true, 00:29:14.773 "seek_hole": false, 00:29:14.773 "seek_data": false, 00:29:14.773 "copy": true, 00:29:14.773 "nvme_iov_md": false 00:29:14.773 }, 00:29:14.773 "memory_domains": [ 00:29:14.773 { 00:29:14.773 "dma_device_id": "system", 00:29:14.773 "dma_device_type": 1 00:29:14.773 }, 00:29:14.773 { 00:29:14.773 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:14.773 "dma_device_type": 2 00:29:14.773 } 00:29:14.773 ], 00:29:14.773 "driver_specific": {} 00:29:14.773 } 00:29:14.773 ] 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:14.773 BaseBdev3 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:14.773 [ 00:29:14.773 { 00:29:14.773 "name": "BaseBdev3", 00:29:14.773 "aliases": [ 00:29:14.773 "9a61fc2c-9538-40cb-829f-902d0255a850" 00:29:14.773 ], 00:29:14.773 "product_name": "Malloc disk", 00:29:14.773 "block_size": 512, 00:29:14.773 "num_blocks": 65536, 00:29:14.773 "uuid": "9a61fc2c-9538-40cb-829f-902d0255a850", 00:29:14.773 "assigned_rate_limits": { 00:29:14.773 "rw_ios_per_sec": 0, 00:29:14.773 "rw_mbytes_per_sec": 0, 00:29:14.773 "r_mbytes_per_sec": 0, 00:29:14.773 "w_mbytes_per_sec": 0 00:29:14.773 }, 00:29:14.773 "claimed": false, 00:29:14.773 "zoned": false, 00:29:14.773 "supported_io_types": { 00:29:14.773 "read": true, 00:29:14.773 "write": true, 00:29:14.773 "unmap": true, 00:29:14.773 "flush": true, 00:29:14.773 "reset": true, 00:29:14.773 "nvme_admin": false, 00:29:14.773 "nvme_io": false, 00:29:14.773 "nvme_io_md": false, 00:29:14.773 "write_zeroes": true, 00:29:14.773 "zcopy": true, 00:29:14.773 "get_zone_info": false, 00:29:14.773 "zone_management": false, 00:29:14.773 "zone_append": false, 00:29:14.773 "compare": false, 00:29:14.773 "compare_and_write": false, 00:29:14.773 "abort": true, 00:29:14.773 "seek_hole": false, 00:29:14.773 "seek_data": false, 00:29:14.773 "copy": true, 00:29:14.773 "nvme_iov_md": false 00:29:14.773 }, 00:29:14.773 "memory_domains": [ 00:29:14.773 { 00:29:14.773 "dma_device_id": "system", 00:29:14.773 "dma_device_type": 1 00:29:14.773 }, 00:29:14.773 { 00:29:14.773 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:14.773 "dma_device_type": 2 00:29:14.773 } 00:29:14.773 ], 00:29:14.773 "driver_specific": {} 00:29:14.773 } 00:29:14.773 ] 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:14.773 [2024-10-01 20:28:09.904673] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:29:14.773 [2024-10-01 20:28:09.904801] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:29:14.773 [2024-10-01 20:28:09.904892] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:14.773 [2024-10-01 20:28:09.908877] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:14.773 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:14.773 "name": "Existed_Raid", 00:29:14.773 "uuid": "f1ee8ac5-1149-4962-82bd-d986bbaa86a2", 00:29:14.773 "strip_size_kb": 64, 00:29:14.773 "state": "configuring", 00:29:14.773 "raid_level": "raid5f", 00:29:14.773 "superblock": true, 00:29:14.773 "num_base_bdevs": 3, 00:29:14.773 "num_base_bdevs_discovered": 2, 00:29:14.773 "num_base_bdevs_operational": 3, 00:29:14.773 "base_bdevs_list": [ 00:29:14.773 { 00:29:14.773 "name": "BaseBdev1", 00:29:14.773 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:14.773 "is_configured": false, 00:29:14.773 "data_offset": 0, 00:29:14.773 "data_size": 0 00:29:14.773 }, 00:29:14.773 { 00:29:14.773 "name": "BaseBdev2", 00:29:14.773 "uuid": "8184849d-05c0-447c-b4fe-01784bc12d3f", 00:29:14.773 "is_configured": true, 00:29:14.773 "data_offset": 2048, 00:29:14.773 "data_size": 63488 00:29:14.773 }, 00:29:14.773 { 00:29:14.773 "name": "BaseBdev3", 00:29:14.773 "uuid": "9a61fc2c-9538-40cb-829f-902d0255a850", 00:29:14.773 "is_configured": true, 00:29:14.773 "data_offset": 2048, 00:29:14.773 "data_size": 63488 00:29:14.773 } 00:29:14.774 ] 00:29:14.774 }' 00:29:14.774 20:28:09 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:14.774 20:28:09 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:15.341 [2024-10-01 20:28:10.445419] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:15.341 "name": "Existed_Raid", 00:29:15.341 "uuid": "f1ee8ac5-1149-4962-82bd-d986bbaa86a2", 00:29:15.341 "strip_size_kb": 64, 00:29:15.341 "state": "configuring", 00:29:15.341 "raid_level": "raid5f", 00:29:15.341 "superblock": true, 00:29:15.341 "num_base_bdevs": 3, 00:29:15.341 "num_base_bdevs_discovered": 1, 00:29:15.341 "num_base_bdevs_operational": 3, 00:29:15.341 "base_bdevs_list": [ 00:29:15.341 { 00:29:15.341 "name": "BaseBdev1", 00:29:15.341 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:15.341 "is_configured": false, 00:29:15.341 "data_offset": 0, 00:29:15.341 "data_size": 0 00:29:15.341 }, 00:29:15.341 { 00:29:15.341 "name": null, 00:29:15.341 "uuid": "8184849d-05c0-447c-b4fe-01784bc12d3f", 00:29:15.341 "is_configured": false, 00:29:15.341 "data_offset": 0, 00:29:15.341 "data_size": 63488 00:29:15.341 }, 00:29:15.341 { 00:29:15.341 "name": "BaseBdev3", 00:29:15.341 "uuid": "9a61fc2c-9538-40cb-829f-902d0255a850", 00:29:15.341 "is_configured": true, 00:29:15.341 "data_offset": 2048, 00:29:15.341 "data_size": 63488 00:29:15.341 } 00:29:15.341 ] 00:29:15.341 }' 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:15.341 20:28:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:15.907 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:29:15.907 20:28:10 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:15.907 20:28:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:15.907 20:28:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:15.907 20:28:10 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:15.907 [2024-10-01 20:28:11.045433] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:15.907 BaseBdev1 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:15.907 [ 00:29:15.907 { 00:29:15.907 "name": "BaseBdev1", 00:29:15.907 "aliases": [ 00:29:15.907 "95d6f613-9c0b-49d2-b051-f04691a3a705" 00:29:15.907 ], 00:29:15.907 "product_name": "Malloc disk", 00:29:15.907 "block_size": 512, 00:29:15.907 "num_blocks": 65536, 00:29:15.907 "uuid": "95d6f613-9c0b-49d2-b051-f04691a3a705", 00:29:15.907 "assigned_rate_limits": { 00:29:15.907 "rw_ios_per_sec": 0, 00:29:15.907 "rw_mbytes_per_sec": 0, 00:29:15.907 "r_mbytes_per_sec": 0, 00:29:15.907 "w_mbytes_per_sec": 0 00:29:15.907 }, 00:29:15.907 "claimed": true, 00:29:15.907 "claim_type": "exclusive_write", 00:29:15.907 "zoned": false, 00:29:15.907 "supported_io_types": { 00:29:15.907 "read": true, 00:29:15.907 "write": true, 00:29:15.907 "unmap": true, 00:29:15.907 "flush": true, 00:29:15.907 "reset": true, 00:29:15.907 "nvme_admin": false, 00:29:15.907 "nvme_io": false, 00:29:15.907 "nvme_io_md": false, 00:29:15.907 "write_zeroes": true, 00:29:15.907 "zcopy": true, 00:29:15.907 "get_zone_info": false, 00:29:15.907 "zone_management": false, 00:29:15.907 "zone_append": false, 00:29:15.907 "compare": false, 00:29:15.907 "compare_and_write": false, 00:29:15.907 "abort": true, 00:29:15.907 "seek_hole": false, 00:29:15.907 "seek_data": false, 00:29:15.907 "copy": true, 00:29:15.907 "nvme_iov_md": false 00:29:15.907 }, 00:29:15.907 "memory_domains": [ 00:29:15.907 { 00:29:15.907 "dma_device_id": "system", 00:29:15.907 "dma_device_type": 1 00:29:15.907 }, 00:29:15.907 { 00:29:15.907 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:15.907 "dma_device_type": 2 00:29:15.907 } 00:29:15.907 ], 00:29:15.907 "driver_specific": {} 00:29:15.907 } 00:29:15.907 ] 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:15.907 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:15.907 "name": "Existed_Raid", 00:29:15.907 "uuid": "f1ee8ac5-1149-4962-82bd-d986bbaa86a2", 00:29:15.908 "strip_size_kb": 64, 00:29:15.908 "state": "configuring", 00:29:15.908 "raid_level": "raid5f", 00:29:15.908 "superblock": true, 00:29:15.908 "num_base_bdevs": 3, 00:29:15.908 "num_base_bdevs_discovered": 2, 00:29:15.908 "num_base_bdevs_operational": 3, 00:29:15.908 "base_bdevs_list": [ 00:29:15.908 { 00:29:15.908 "name": "BaseBdev1", 00:29:15.908 "uuid": "95d6f613-9c0b-49d2-b051-f04691a3a705", 00:29:15.908 "is_configured": true, 00:29:15.908 "data_offset": 2048, 00:29:15.908 "data_size": 63488 00:29:15.908 }, 00:29:15.908 { 00:29:15.908 "name": null, 00:29:15.908 "uuid": "8184849d-05c0-447c-b4fe-01784bc12d3f", 00:29:15.908 "is_configured": false, 00:29:15.908 "data_offset": 0, 00:29:15.908 "data_size": 63488 00:29:15.908 }, 00:29:15.908 { 00:29:15.908 "name": "BaseBdev3", 00:29:15.908 "uuid": "9a61fc2c-9538-40cb-829f-902d0255a850", 00:29:15.908 "is_configured": true, 00:29:15.908 "data_offset": 2048, 00:29:15.908 "data_size": 63488 00:29:15.908 } 00:29:15.908 ] 00:29:15.908 }' 00:29:15.908 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:15.908 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:16.476 [2024-10-01 20:28:11.665684] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:16.476 "name": "Existed_Raid", 00:29:16.476 "uuid": "f1ee8ac5-1149-4962-82bd-d986bbaa86a2", 00:29:16.476 "strip_size_kb": 64, 00:29:16.476 "state": "configuring", 00:29:16.476 "raid_level": "raid5f", 00:29:16.476 "superblock": true, 00:29:16.476 "num_base_bdevs": 3, 00:29:16.476 "num_base_bdevs_discovered": 1, 00:29:16.476 "num_base_bdevs_operational": 3, 00:29:16.476 "base_bdevs_list": [ 00:29:16.476 { 00:29:16.476 "name": "BaseBdev1", 00:29:16.476 "uuid": "95d6f613-9c0b-49d2-b051-f04691a3a705", 00:29:16.476 "is_configured": true, 00:29:16.476 "data_offset": 2048, 00:29:16.476 "data_size": 63488 00:29:16.476 }, 00:29:16.476 { 00:29:16.476 "name": null, 00:29:16.476 "uuid": "8184849d-05c0-447c-b4fe-01784bc12d3f", 00:29:16.476 "is_configured": false, 00:29:16.476 "data_offset": 0, 00:29:16.476 "data_size": 63488 00:29:16.476 }, 00:29:16.476 { 00:29:16.476 "name": null, 00:29:16.476 "uuid": "9a61fc2c-9538-40cb-829f-902d0255a850", 00:29:16.476 "is_configured": false, 00:29:16.476 "data_offset": 0, 00:29:16.476 "data_size": 63488 00:29:16.476 } 00:29:16.476 ] 00:29:16.476 }' 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:16.476 20:28:11 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:17.046 [2024-10-01 20:28:12.257837] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:17.046 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:17.047 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:17.047 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:17.047 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:17.047 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:17.047 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:17.306 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:17.306 "name": "Existed_Raid", 00:29:17.306 "uuid": "f1ee8ac5-1149-4962-82bd-d986bbaa86a2", 00:29:17.306 "strip_size_kb": 64, 00:29:17.306 "state": "configuring", 00:29:17.306 "raid_level": "raid5f", 00:29:17.306 "superblock": true, 00:29:17.306 "num_base_bdevs": 3, 00:29:17.306 "num_base_bdevs_discovered": 2, 00:29:17.306 "num_base_bdevs_operational": 3, 00:29:17.306 "base_bdevs_list": [ 00:29:17.306 { 00:29:17.306 "name": "BaseBdev1", 00:29:17.306 "uuid": "95d6f613-9c0b-49d2-b051-f04691a3a705", 00:29:17.306 "is_configured": true, 00:29:17.307 "data_offset": 2048, 00:29:17.307 "data_size": 63488 00:29:17.307 }, 00:29:17.307 { 00:29:17.307 "name": null, 00:29:17.307 "uuid": "8184849d-05c0-447c-b4fe-01784bc12d3f", 00:29:17.307 "is_configured": false, 00:29:17.307 "data_offset": 0, 00:29:17.307 "data_size": 63488 00:29:17.307 }, 00:29:17.307 { 00:29:17.307 "name": "BaseBdev3", 00:29:17.307 "uuid": "9a61fc2c-9538-40cb-829f-902d0255a850", 00:29:17.307 "is_configured": true, 00:29:17.307 "data_offset": 2048, 00:29:17.307 "data_size": 63488 00:29:17.307 } 00:29:17.307 ] 00:29:17.307 }' 00:29:17.307 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:17.307 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:17.566 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:29:17.566 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:17.566 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:17.566 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:17.566 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:17.825 [2024-10-01 20:28:12.826128] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:17.825 "name": "Existed_Raid", 00:29:17.825 "uuid": "f1ee8ac5-1149-4962-82bd-d986bbaa86a2", 00:29:17.825 "strip_size_kb": 64, 00:29:17.825 "state": "configuring", 00:29:17.825 "raid_level": "raid5f", 00:29:17.825 "superblock": true, 00:29:17.825 "num_base_bdevs": 3, 00:29:17.825 "num_base_bdevs_discovered": 1, 00:29:17.825 "num_base_bdevs_operational": 3, 00:29:17.825 "base_bdevs_list": [ 00:29:17.825 { 00:29:17.825 "name": null, 00:29:17.825 "uuid": "95d6f613-9c0b-49d2-b051-f04691a3a705", 00:29:17.825 "is_configured": false, 00:29:17.825 "data_offset": 0, 00:29:17.825 "data_size": 63488 00:29:17.825 }, 00:29:17.825 { 00:29:17.825 "name": null, 00:29:17.825 "uuid": "8184849d-05c0-447c-b4fe-01784bc12d3f", 00:29:17.825 "is_configured": false, 00:29:17.825 "data_offset": 0, 00:29:17.825 "data_size": 63488 00:29:17.825 }, 00:29:17.825 { 00:29:17.825 "name": "BaseBdev3", 00:29:17.825 "uuid": "9a61fc2c-9538-40cb-829f-902d0255a850", 00:29:17.825 "is_configured": true, 00:29:17.825 "data_offset": 2048, 00:29:17.825 "data_size": 63488 00:29:17.825 } 00:29:17.825 ] 00:29:17.825 }' 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:17.825 20:28:12 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:18.394 [2024-10-01 20:28:13.513674] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:18.394 "name": "Existed_Raid", 00:29:18.394 "uuid": "f1ee8ac5-1149-4962-82bd-d986bbaa86a2", 00:29:18.394 "strip_size_kb": 64, 00:29:18.394 "state": "configuring", 00:29:18.394 "raid_level": "raid5f", 00:29:18.394 "superblock": true, 00:29:18.394 "num_base_bdevs": 3, 00:29:18.394 "num_base_bdevs_discovered": 2, 00:29:18.394 "num_base_bdevs_operational": 3, 00:29:18.394 "base_bdevs_list": [ 00:29:18.394 { 00:29:18.394 "name": null, 00:29:18.394 "uuid": "95d6f613-9c0b-49d2-b051-f04691a3a705", 00:29:18.394 "is_configured": false, 00:29:18.394 "data_offset": 0, 00:29:18.394 "data_size": 63488 00:29:18.394 }, 00:29:18.394 { 00:29:18.394 "name": "BaseBdev2", 00:29:18.394 "uuid": "8184849d-05c0-447c-b4fe-01784bc12d3f", 00:29:18.394 "is_configured": true, 00:29:18.394 "data_offset": 2048, 00:29:18.394 "data_size": 63488 00:29:18.394 }, 00:29:18.394 { 00:29:18.394 "name": "BaseBdev3", 00:29:18.394 "uuid": "9a61fc2c-9538-40cb-829f-902d0255a850", 00:29:18.394 "is_configured": true, 00:29:18.394 "data_offset": 2048, 00:29:18.394 "data_size": 63488 00:29:18.394 } 00:29:18.394 ] 00:29:18.394 }' 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:18.394 20:28:13 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 95d6f613-9c0b-49d2-b051-f04691a3a705 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:19.020 [2024-10-01 20:28:14.176500] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:29:19.020 NewBaseBdev 00:29:19.020 [2024-10-01 20:28:14.177000] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:29:19.020 [2024-10-01 20:28:14.177036] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:29:19.020 [2024-10-01 20:28:14.177379] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:19.020 [2024-10-01 20:28:14.182433] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:29:19.020 [2024-10-01 20:28:14.182589] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:29:19.020 [2024-10-01 20:28:14.182952] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:19.020 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:19.020 [ 00:29:19.020 { 00:29:19.020 "name": "NewBaseBdev", 00:29:19.020 "aliases": [ 00:29:19.020 "95d6f613-9c0b-49d2-b051-f04691a3a705" 00:29:19.020 ], 00:29:19.020 "product_name": "Malloc disk", 00:29:19.020 "block_size": 512, 00:29:19.020 "num_blocks": 65536, 00:29:19.020 "uuid": "95d6f613-9c0b-49d2-b051-f04691a3a705", 00:29:19.020 "assigned_rate_limits": { 00:29:19.020 "rw_ios_per_sec": 0, 00:29:19.020 "rw_mbytes_per_sec": 0, 00:29:19.020 "r_mbytes_per_sec": 0, 00:29:19.020 "w_mbytes_per_sec": 0 00:29:19.020 }, 00:29:19.020 "claimed": true, 00:29:19.020 "claim_type": "exclusive_write", 00:29:19.020 "zoned": false, 00:29:19.020 "supported_io_types": { 00:29:19.020 "read": true, 00:29:19.020 "write": true, 00:29:19.020 "unmap": true, 00:29:19.020 "flush": true, 00:29:19.020 "reset": true, 00:29:19.020 "nvme_admin": false, 00:29:19.020 "nvme_io": false, 00:29:19.020 "nvme_io_md": false, 00:29:19.020 "write_zeroes": true, 00:29:19.020 "zcopy": true, 00:29:19.020 "get_zone_info": false, 00:29:19.020 "zone_management": false, 00:29:19.020 "zone_append": false, 00:29:19.020 "compare": false, 00:29:19.020 "compare_and_write": false, 00:29:19.020 "abort": true, 00:29:19.020 "seek_hole": false, 00:29:19.020 "seek_data": false, 00:29:19.020 "copy": true, 00:29:19.020 "nvme_iov_md": false 00:29:19.020 }, 00:29:19.020 "memory_domains": [ 00:29:19.020 { 00:29:19.020 "dma_device_id": "system", 00:29:19.020 "dma_device_type": 1 00:29:19.020 }, 00:29:19.020 { 00:29:19.020 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:29:19.020 "dma_device_type": 2 00:29:19.020 } 00:29:19.020 ], 00:29:19.020 "driver_specific": {} 00:29:19.020 } 00:29:19.020 ] 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:19.021 "name": "Existed_Raid", 00:29:19.021 "uuid": "f1ee8ac5-1149-4962-82bd-d986bbaa86a2", 00:29:19.021 "strip_size_kb": 64, 00:29:19.021 "state": "online", 00:29:19.021 "raid_level": "raid5f", 00:29:19.021 "superblock": true, 00:29:19.021 "num_base_bdevs": 3, 00:29:19.021 "num_base_bdevs_discovered": 3, 00:29:19.021 "num_base_bdevs_operational": 3, 00:29:19.021 "base_bdevs_list": [ 00:29:19.021 { 00:29:19.021 "name": "NewBaseBdev", 00:29:19.021 "uuid": "95d6f613-9c0b-49d2-b051-f04691a3a705", 00:29:19.021 "is_configured": true, 00:29:19.021 "data_offset": 2048, 00:29:19.021 "data_size": 63488 00:29:19.021 }, 00:29:19.021 { 00:29:19.021 "name": "BaseBdev2", 00:29:19.021 "uuid": "8184849d-05c0-447c-b4fe-01784bc12d3f", 00:29:19.021 "is_configured": true, 00:29:19.021 "data_offset": 2048, 00:29:19.021 "data_size": 63488 00:29:19.021 }, 00:29:19.021 { 00:29:19.021 "name": "BaseBdev3", 00:29:19.021 "uuid": "9a61fc2c-9538-40cb-829f-902d0255a850", 00:29:19.021 "is_configured": true, 00:29:19.021 "data_offset": 2048, 00:29:19.021 "data_size": 63488 00:29:19.021 } 00:29:19.021 ] 00:29:19.021 }' 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:19.021 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:19.590 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:29:19.590 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:29:19.590 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:19.590 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:19.590 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:29:19.590 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:19.590 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:19.590 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:29:19.590 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:19.590 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:19.590 [2024-10-01 20:28:14.736998] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:19.590 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:19.590 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:19.590 "name": "Existed_Raid", 00:29:19.590 "aliases": [ 00:29:19.590 "f1ee8ac5-1149-4962-82bd-d986bbaa86a2" 00:29:19.590 ], 00:29:19.590 "product_name": "Raid Volume", 00:29:19.590 "block_size": 512, 00:29:19.590 "num_blocks": 126976, 00:29:19.590 "uuid": "f1ee8ac5-1149-4962-82bd-d986bbaa86a2", 00:29:19.590 "assigned_rate_limits": { 00:29:19.590 "rw_ios_per_sec": 0, 00:29:19.590 "rw_mbytes_per_sec": 0, 00:29:19.590 "r_mbytes_per_sec": 0, 00:29:19.590 "w_mbytes_per_sec": 0 00:29:19.590 }, 00:29:19.590 "claimed": false, 00:29:19.590 "zoned": false, 00:29:19.590 "supported_io_types": { 00:29:19.590 "read": true, 00:29:19.590 "write": true, 00:29:19.590 "unmap": false, 00:29:19.590 "flush": false, 00:29:19.590 "reset": true, 00:29:19.590 "nvme_admin": false, 00:29:19.590 "nvme_io": false, 00:29:19.590 "nvme_io_md": false, 00:29:19.590 "write_zeroes": true, 00:29:19.590 "zcopy": false, 00:29:19.590 "get_zone_info": false, 00:29:19.590 "zone_management": false, 00:29:19.590 "zone_append": false, 00:29:19.590 "compare": false, 00:29:19.590 "compare_and_write": false, 00:29:19.590 "abort": false, 00:29:19.590 "seek_hole": false, 00:29:19.590 "seek_data": false, 00:29:19.590 "copy": false, 00:29:19.590 "nvme_iov_md": false 00:29:19.590 }, 00:29:19.590 "driver_specific": { 00:29:19.590 "raid": { 00:29:19.590 "uuid": "f1ee8ac5-1149-4962-82bd-d986bbaa86a2", 00:29:19.590 "strip_size_kb": 64, 00:29:19.590 "state": "online", 00:29:19.590 "raid_level": "raid5f", 00:29:19.590 "superblock": true, 00:29:19.590 "num_base_bdevs": 3, 00:29:19.590 "num_base_bdevs_discovered": 3, 00:29:19.590 "num_base_bdevs_operational": 3, 00:29:19.590 "base_bdevs_list": [ 00:29:19.590 { 00:29:19.590 "name": "NewBaseBdev", 00:29:19.590 "uuid": "95d6f613-9c0b-49d2-b051-f04691a3a705", 00:29:19.590 "is_configured": true, 00:29:19.590 "data_offset": 2048, 00:29:19.590 "data_size": 63488 00:29:19.590 }, 00:29:19.590 { 00:29:19.590 "name": "BaseBdev2", 00:29:19.590 "uuid": "8184849d-05c0-447c-b4fe-01784bc12d3f", 00:29:19.590 "is_configured": true, 00:29:19.590 "data_offset": 2048, 00:29:19.590 "data_size": 63488 00:29:19.590 }, 00:29:19.590 { 00:29:19.590 "name": "BaseBdev3", 00:29:19.590 "uuid": "9a61fc2c-9538-40cb-829f-902d0255a850", 00:29:19.590 "is_configured": true, 00:29:19.590 "data_offset": 2048, 00:29:19.590 "data_size": 63488 00:29:19.590 } 00:29:19.590 ] 00:29:19.590 } 00:29:19.590 } 00:29:19.590 }' 00:29:19.590 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:19.590 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:29:19.590 BaseBdev2 00:29:19.590 BaseBdev3' 00:29:19.590 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:19.851 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:19.851 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:19.851 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:29:19.851 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:19.851 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:19.851 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:19.851 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:19.851 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:19.851 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:19.851 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:19.851 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:29:19.851 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:19.851 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:19.851 20:28:14 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:19.851 20:28:14 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:19.851 20:28:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:19.851 20:28:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:19.852 [2024-10-01 20:28:15.064756] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:29:19.852 [2024-10-01 20:28:15.064935] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:19.852 [2024-10-01 20:28:15.065167] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:19.852 [2024-10-01 20:28:15.065694] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:19.852 [2024-10-01 20:28:15.065863] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 81732 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 81732 ']' 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 81732 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 81732 00:29:19.852 killing process with pid 81732 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 81732' 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 81732 00:29:19.852 [2024-10-01 20:28:15.103617] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:29:19.852 20:28:15 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 81732 00:29:20.421 [2024-10-01 20:28:15.395880] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:29:22.327 20:28:17 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:29:22.327 00:29:22.327 real 0m13.130s 00:29:22.327 user 0m20.861s 00:29:22.327 sys 0m1.880s 00:29:22.327 20:28:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:29:22.327 ************************************ 00:29:22.327 END TEST raid5f_state_function_test_sb 00:29:22.327 ************************************ 00:29:22.327 20:28:17 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:22.327 20:28:17 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 3 00:29:22.327 20:28:17 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:29:22.327 20:28:17 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:29:22.327 20:28:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:29:22.327 ************************************ 00:29:22.327 START TEST raid5f_superblock_test 00:29:22.327 ************************************ 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid5f 3 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=82375 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 82375 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 82375 ']' 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:22.327 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:29:22.327 20:28:17 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:22.327 [2024-10-01 20:28:17.549945] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:29:22.327 [2024-10-01 20:28:17.550387] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid82375 ] 00:29:22.587 [2024-10-01 20:28:17.732082] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:22.846 [2024-10-01 20:28:18.016907] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:29:23.105 [2024-10-01 20:28:18.237199] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:23.105 [2024-10-01 20:28:18.237260] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.673 malloc1 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.673 [2024-10-01 20:28:18.707589] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:29:23.673 [2024-10-01 20:28:18.707832] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:23.673 [2024-10-01 20:28:18.708035] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:29:23.673 [2024-10-01 20:28:18.708181] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:23.673 [2024-10-01 20:28:18.711238] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:23.673 [2024-10-01 20:28:18.711287] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:29:23.673 pt1 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.673 malloc2 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.673 [2024-10-01 20:28:18.764883] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:29:23.673 [2024-10-01 20:28:18.765100] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:23.673 [2024-10-01 20:28:18.765189] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:29:23.673 [2024-10-01 20:28:18.765406] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:23.673 [2024-10-01 20:28:18.768485] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:23.673 [2024-10-01 20:28:18.768653] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:29:23.673 pt2 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:23.673 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.674 malloc3 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.674 [2024-10-01 20:28:18.821780] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:29:23.674 [2024-10-01 20:28:18.821992] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:23.674 [2024-10-01 20:28:18.822039] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:29:23.674 [2024-10-01 20:28:18.822058] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:23.674 [2024-10-01 20:28:18.825386] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:23.674 [2024-10-01 20:28:18.825433] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:29:23.674 pt3 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.674 [2024-10-01 20:28:18.831333] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:29:23.674 [2024-10-01 20:28:18.834350] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:29:23.674 [2024-10-01 20:28:18.834450] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:29:23.674 [2024-10-01 20:28:18.834722] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:29:23.674 [2024-10-01 20:28:18.834805] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:29:23.674 [2024-10-01 20:28:18.835157] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:29:23.674 [2024-10-01 20:28:18.840706] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:29:23.674 [2024-10-01 20:28:18.840890] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:29:23.674 [2024-10-01 20:28:18.841174] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:23.674 "name": "raid_bdev1", 00:29:23.674 "uuid": "8b29190a-dd4f-43da-af19-c2f886a47903", 00:29:23.674 "strip_size_kb": 64, 00:29:23.674 "state": "online", 00:29:23.674 "raid_level": "raid5f", 00:29:23.674 "superblock": true, 00:29:23.674 "num_base_bdevs": 3, 00:29:23.674 "num_base_bdevs_discovered": 3, 00:29:23.674 "num_base_bdevs_operational": 3, 00:29:23.674 "base_bdevs_list": [ 00:29:23.674 { 00:29:23.674 "name": "pt1", 00:29:23.674 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:23.674 "is_configured": true, 00:29:23.674 "data_offset": 2048, 00:29:23.674 "data_size": 63488 00:29:23.674 }, 00:29:23.674 { 00:29:23.674 "name": "pt2", 00:29:23.674 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:23.674 "is_configured": true, 00:29:23.674 "data_offset": 2048, 00:29:23.674 "data_size": 63488 00:29:23.674 }, 00:29:23.674 { 00:29:23.674 "name": "pt3", 00:29:23.674 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:23.674 "is_configured": true, 00:29:23.674 "data_offset": 2048, 00:29:23.674 "data_size": 63488 00:29:23.674 } 00:29:23.674 ] 00:29:23.674 }' 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:23.674 20:28:18 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.244 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:29:24.244 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:29:24.244 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:24.244 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:24.244 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:29:24.244 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:24.244 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:29:24.244 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:24.244 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:24.244 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.244 [2024-10-01 20:28:19.347388] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:24.244 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:24.244 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:24.244 "name": "raid_bdev1", 00:29:24.244 "aliases": [ 00:29:24.244 "8b29190a-dd4f-43da-af19-c2f886a47903" 00:29:24.244 ], 00:29:24.244 "product_name": "Raid Volume", 00:29:24.244 "block_size": 512, 00:29:24.244 "num_blocks": 126976, 00:29:24.244 "uuid": "8b29190a-dd4f-43da-af19-c2f886a47903", 00:29:24.244 "assigned_rate_limits": { 00:29:24.244 "rw_ios_per_sec": 0, 00:29:24.244 "rw_mbytes_per_sec": 0, 00:29:24.244 "r_mbytes_per_sec": 0, 00:29:24.244 "w_mbytes_per_sec": 0 00:29:24.244 }, 00:29:24.244 "claimed": false, 00:29:24.244 "zoned": false, 00:29:24.244 "supported_io_types": { 00:29:24.244 "read": true, 00:29:24.244 "write": true, 00:29:24.244 "unmap": false, 00:29:24.244 "flush": false, 00:29:24.244 "reset": true, 00:29:24.244 "nvme_admin": false, 00:29:24.244 "nvme_io": false, 00:29:24.244 "nvme_io_md": false, 00:29:24.244 "write_zeroes": true, 00:29:24.244 "zcopy": false, 00:29:24.244 "get_zone_info": false, 00:29:24.244 "zone_management": false, 00:29:24.244 "zone_append": false, 00:29:24.244 "compare": false, 00:29:24.244 "compare_and_write": false, 00:29:24.244 "abort": false, 00:29:24.244 "seek_hole": false, 00:29:24.244 "seek_data": false, 00:29:24.244 "copy": false, 00:29:24.244 "nvme_iov_md": false 00:29:24.244 }, 00:29:24.244 "driver_specific": { 00:29:24.244 "raid": { 00:29:24.244 "uuid": "8b29190a-dd4f-43da-af19-c2f886a47903", 00:29:24.244 "strip_size_kb": 64, 00:29:24.244 "state": "online", 00:29:24.244 "raid_level": "raid5f", 00:29:24.244 "superblock": true, 00:29:24.244 "num_base_bdevs": 3, 00:29:24.244 "num_base_bdevs_discovered": 3, 00:29:24.244 "num_base_bdevs_operational": 3, 00:29:24.244 "base_bdevs_list": [ 00:29:24.244 { 00:29:24.244 "name": "pt1", 00:29:24.244 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:24.244 "is_configured": true, 00:29:24.244 "data_offset": 2048, 00:29:24.244 "data_size": 63488 00:29:24.244 }, 00:29:24.244 { 00:29:24.244 "name": "pt2", 00:29:24.244 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:24.244 "is_configured": true, 00:29:24.244 "data_offset": 2048, 00:29:24.244 "data_size": 63488 00:29:24.244 }, 00:29:24.244 { 00:29:24.244 "name": "pt3", 00:29:24.244 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:24.244 "is_configured": true, 00:29:24.244 "data_offset": 2048, 00:29:24.244 "data_size": 63488 00:29:24.244 } 00:29:24.244 ] 00:29:24.244 } 00:29:24.244 } 00:29:24.244 }' 00:29:24.244 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:24.244 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:29:24.244 pt2 00:29:24.244 pt3' 00:29:24.244 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:29:24.504 [2024-10-01 20:28:19.663426] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=8b29190a-dd4f-43da-af19-c2f886a47903 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 8b29190a-dd4f-43da-af19-c2f886a47903 ']' 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.504 [2024-10-01 20:28:19.715188] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:29:24.504 [2024-10-01 20:28:19.715343] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:24.504 [2024-10-01 20:28:19.715574] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:24.504 [2024-10-01 20:28:19.715836] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:24.504 [2024-10-01 20:28:19.715987] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.504 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:24.763 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.764 [2024-10-01 20:28:19.867272] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:29:24.764 [2024-10-01 20:28:19.870114] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:29:24.764 [2024-10-01 20:28:19.870276] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:29:24.764 [2024-10-01 20:28:19.870373] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:29:24.764 [2024-10-01 20:28:19.870449] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:29:24.764 [2024-10-01 20:28:19.870486] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:29:24.764 [2024-10-01 20:28:19.870517] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:29:24.764 [2024-10-01 20:28:19.870534] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:29:24.764 request: 00:29:24.764 { 00:29:24.764 "name": "raid_bdev1", 00:29:24.764 "raid_level": "raid5f", 00:29:24.764 "base_bdevs": [ 00:29:24.764 "malloc1", 00:29:24.764 "malloc2", 00:29:24.764 "malloc3" 00:29:24.764 ], 00:29:24.764 "strip_size_kb": 64, 00:29:24.764 "superblock": false, 00:29:24.764 "method": "bdev_raid_create", 00:29:24.764 "req_id": 1 00:29:24.764 } 00:29:24.764 Got JSON-RPC error response 00:29:24.764 response: 00:29:24.764 { 00:29:24.764 "code": -17, 00:29:24.764 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:29:24.764 } 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.764 [2024-10-01 20:28:19.935268] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:29:24.764 [2024-10-01 20:28:19.935463] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:24.764 [2024-10-01 20:28:19.935551] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:29:24.764 [2024-10-01 20:28:19.935706] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:24.764 [2024-10-01 20:28:19.938992] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:24.764 [2024-10-01 20:28:19.939150] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:29:24.764 [2024-10-01 20:28:19.939358] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:29:24.764 [2024-10-01 20:28:19.939552] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:29:24.764 pt1 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:24.764 "name": "raid_bdev1", 00:29:24.764 "uuid": "8b29190a-dd4f-43da-af19-c2f886a47903", 00:29:24.764 "strip_size_kb": 64, 00:29:24.764 "state": "configuring", 00:29:24.764 "raid_level": "raid5f", 00:29:24.764 "superblock": true, 00:29:24.764 "num_base_bdevs": 3, 00:29:24.764 "num_base_bdevs_discovered": 1, 00:29:24.764 "num_base_bdevs_operational": 3, 00:29:24.764 "base_bdevs_list": [ 00:29:24.764 { 00:29:24.764 "name": "pt1", 00:29:24.764 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:24.764 "is_configured": true, 00:29:24.764 "data_offset": 2048, 00:29:24.764 "data_size": 63488 00:29:24.764 }, 00:29:24.764 { 00:29:24.764 "name": null, 00:29:24.764 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:24.764 "is_configured": false, 00:29:24.764 "data_offset": 2048, 00:29:24.764 "data_size": 63488 00:29:24.764 }, 00:29:24.764 { 00:29:24.764 "name": null, 00:29:24.764 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:24.764 "is_configured": false, 00:29:24.764 "data_offset": 2048, 00:29:24.764 "data_size": 63488 00:29:24.764 } 00:29:24.764 ] 00:29:24.764 }' 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:24.764 20:28:19 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.333 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:29:25.333 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:29:25.333 20:28:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:25.333 20:28:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.333 [2024-10-01 20:28:20.471632] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:29:25.333 [2024-10-01 20:28:20.471888] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:25.333 [2024-10-01 20:28:20.471950] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:29:25.333 [2024-10-01 20:28:20.471975] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:25.333 [2024-10-01 20:28:20.472602] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:25.333 [2024-10-01 20:28:20.472644] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:29:25.333 [2024-10-01 20:28:20.472784] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:29:25.333 [2024-10-01 20:28:20.472820] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:29:25.333 pt2 00:29:25.333 20:28:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:25.333 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:29:25.333 20:28:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:25.333 20:28:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.333 [2024-10-01 20:28:20.479630] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:29:25.333 20:28:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:25.333 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:25.334 "name": "raid_bdev1", 00:29:25.334 "uuid": "8b29190a-dd4f-43da-af19-c2f886a47903", 00:29:25.334 "strip_size_kb": 64, 00:29:25.334 "state": "configuring", 00:29:25.334 "raid_level": "raid5f", 00:29:25.334 "superblock": true, 00:29:25.334 "num_base_bdevs": 3, 00:29:25.334 "num_base_bdevs_discovered": 1, 00:29:25.334 "num_base_bdevs_operational": 3, 00:29:25.334 "base_bdevs_list": [ 00:29:25.334 { 00:29:25.334 "name": "pt1", 00:29:25.334 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:25.334 "is_configured": true, 00:29:25.334 "data_offset": 2048, 00:29:25.334 "data_size": 63488 00:29:25.334 }, 00:29:25.334 { 00:29:25.334 "name": null, 00:29:25.334 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:25.334 "is_configured": false, 00:29:25.334 "data_offset": 0, 00:29:25.334 "data_size": 63488 00:29:25.334 }, 00:29:25.334 { 00:29:25.334 "name": null, 00:29:25.334 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:25.334 "is_configured": false, 00:29:25.334 "data_offset": 2048, 00:29:25.334 "data_size": 63488 00:29:25.334 } 00:29:25.334 ] 00:29:25.334 }' 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:25.334 20:28:20 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.902 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:29:25.902 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:29:25.902 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:29:25.902 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:25.902 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.902 [2024-10-01 20:28:21.023784] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:29:25.902 [2024-10-01 20:28:21.024039] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:25.902 [2024-10-01 20:28:21.024190] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:29:25.902 [2024-10-01 20:28:21.024327] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:25.902 [2024-10-01 20:28:21.024968] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:25.902 [2024-10-01 20:28:21.025002] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:29:25.902 [2024-10-01 20:28:21.025122] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:29:25.902 [2024-10-01 20:28:21.025162] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:29:25.902 pt2 00:29:25.902 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:25.902 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:29:25.902 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:29:25.902 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:29:25.902 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:25.902 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.902 [2024-10-01 20:28:21.031773] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:29:25.903 [2024-10-01 20:28:21.031975] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:25.903 [2024-10-01 20:28:21.032143] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:29:25.903 [2024-10-01 20:28:21.032286] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:25.903 [2024-10-01 20:28:21.032877] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:25.903 [2024-10-01 20:28:21.033047] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:29:25.903 [2024-10-01 20:28:21.033286] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:29:25.903 [2024-10-01 20:28:21.033478] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:29:25.903 [2024-10-01 20:28:21.033794] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:29:25.903 [2024-10-01 20:28:21.033943] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:29:25.903 [2024-10-01 20:28:21.034281] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:29:25.903 pt3 00:29:25.903 [2024-10-01 20:28:21.039291] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:29:25.903 [2024-10-01 20:28:21.039319] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:29:25.903 [2024-10-01 20:28:21.039559] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:25.903 "name": "raid_bdev1", 00:29:25.903 "uuid": "8b29190a-dd4f-43da-af19-c2f886a47903", 00:29:25.903 "strip_size_kb": 64, 00:29:25.903 "state": "online", 00:29:25.903 "raid_level": "raid5f", 00:29:25.903 "superblock": true, 00:29:25.903 "num_base_bdevs": 3, 00:29:25.903 "num_base_bdevs_discovered": 3, 00:29:25.903 "num_base_bdevs_operational": 3, 00:29:25.903 "base_bdevs_list": [ 00:29:25.903 { 00:29:25.903 "name": "pt1", 00:29:25.903 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:25.903 "is_configured": true, 00:29:25.903 "data_offset": 2048, 00:29:25.903 "data_size": 63488 00:29:25.903 }, 00:29:25.903 { 00:29:25.903 "name": "pt2", 00:29:25.903 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:25.903 "is_configured": true, 00:29:25.903 "data_offset": 2048, 00:29:25.903 "data_size": 63488 00:29:25.903 }, 00:29:25.903 { 00:29:25.903 "name": "pt3", 00:29:25.903 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:25.903 "is_configured": true, 00:29:25.903 "data_offset": 2048, 00:29:25.903 "data_size": 63488 00:29:25.903 } 00:29:25.903 ] 00:29:25.903 }' 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:25.903 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.471 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:29:26.471 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:29:26.471 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:29:26.471 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:29:26.471 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:29:26.471 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:29:26.471 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:29:26.471 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:26.471 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.471 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:29:26.471 [2024-10-01 20:28:21.585956] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:26.471 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:26.471 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:29:26.471 "name": "raid_bdev1", 00:29:26.471 "aliases": [ 00:29:26.471 "8b29190a-dd4f-43da-af19-c2f886a47903" 00:29:26.471 ], 00:29:26.471 "product_name": "Raid Volume", 00:29:26.471 "block_size": 512, 00:29:26.471 "num_blocks": 126976, 00:29:26.471 "uuid": "8b29190a-dd4f-43da-af19-c2f886a47903", 00:29:26.471 "assigned_rate_limits": { 00:29:26.471 "rw_ios_per_sec": 0, 00:29:26.471 "rw_mbytes_per_sec": 0, 00:29:26.471 "r_mbytes_per_sec": 0, 00:29:26.471 "w_mbytes_per_sec": 0 00:29:26.471 }, 00:29:26.471 "claimed": false, 00:29:26.471 "zoned": false, 00:29:26.471 "supported_io_types": { 00:29:26.471 "read": true, 00:29:26.471 "write": true, 00:29:26.471 "unmap": false, 00:29:26.471 "flush": false, 00:29:26.471 "reset": true, 00:29:26.471 "nvme_admin": false, 00:29:26.471 "nvme_io": false, 00:29:26.471 "nvme_io_md": false, 00:29:26.471 "write_zeroes": true, 00:29:26.471 "zcopy": false, 00:29:26.471 "get_zone_info": false, 00:29:26.471 "zone_management": false, 00:29:26.471 "zone_append": false, 00:29:26.471 "compare": false, 00:29:26.471 "compare_and_write": false, 00:29:26.471 "abort": false, 00:29:26.471 "seek_hole": false, 00:29:26.472 "seek_data": false, 00:29:26.472 "copy": false, 00:29:26.472 "nvme_iov_md": false 00:29:26.472 }, 00:29:26.472 "driver_specific": { 00:29:26.472 "raid": { 00:29:26.472 "uuid": "8b29190a-dd4f-43da-af19-c2f886a47903", 00:29:26.472 "strip_size_kb": 64, 00:29:26.472 "state": "online", 00:29:26.472 "raid_level": "raid5f", 00:29:26.472 "superblock": true, 00:29:26.472 "num_base_bdevs": 3, 00:29:26.472 "num_base_bdevs_discovered": 3, 00:29:26.472 "num_base_bdevs_operational": 3, 00:29:26.472 "base_bdevs_list": [ 00:29:26.472 { 00:29:26.472 "name": "pt1", 00:29:26.472 "uuid": "00000000-0000-0000-0000-000000000001", 00:29:26.472 "is_configured": true, 00:29:26.472 "data_offset": 2048, 00:29:26.472 "data_size": 63488 00:29:26.472 }, 00:29:26.472 { 00:29:26.472 "name": "pt2", 00:29:26.472 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:26.472 "is_configured": true, 00:29:26.472 "data_offset": 2048, 00:29:26.472 "data_size": 63488 00:29:26.472 }, 00:29:26.472 { 00:29:26.472 "name": "pt3", 00:29:26.472 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:26.472 "is_configured": true, 00:29:26.472 "data_offset": 2048, 00:29:26.472 "data_size": 63488 00:29:26.472 } 00:29:26.472 ] 00:29:26.472 } 00:29:26.472 } 00:29:26.472 }' 00:29:26.472 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:29:26.472 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:29:26.472 pt2 00:29:26.472 pt3' 00:29:26.472 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.731 [2024-10-01 20:28:21.913908] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 8b29190a-dd4f-43da-af19-c2f886a47903 '!=' 8b29190a-dd4f-43da-af19-c2f886a47903 ']' 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.731 [2024-10-01 20:28:21.965771] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:26.731 20:28:21 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:26.990 20:28:21 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:26.990 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:26.990 "name": "raid_bdev1", 00:29:26.990 "uuid": "8b29190a-dd4f-43da-af19-c2f886a47903", 00:29:26.990 "strip_size_kb": 64, 00:29:26.990 "state": "online", 00:29:26.990 "raid_level": "raid5f", 00:29:26.990 "superblock": true, 00:29:26.990 "num_base_bdevs": 3, 00:29:26.990 "num_base_bdevs_discovered": 2, 00:29:26.990 "num_base_bdevs_operational": 2, 00:29:26.990 "base_bdevs_list": [ 00:29:26.990 { 00:29:26.990 "name": null, 00:29:26.990 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:26.990 "is_configured": false, 00:29:26.990 "data_offset": 0, 00:29:26.990 "data_size": 63488 00:29:26.990 }, 00:29:26.990 { 00:29:26.990 "name": "pt2", 00:29:26.990 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:26.990 "is_configured": true, 00:29:26.990 "data_offset": 2048, 00:29:26.990 "data_size": 63488 00:29:26.990 }, 00:29:26.990 { 00:29:26.990 "name": "pt3", 00:29:26.990 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:26.990 "is_configured": true, 00:29:26.990 "data_offset": 2048, 00:29:26.990 "data_size": 63488 00:29:26.990 } 00:29:26.990 ] 00:29:26.990 }' 00:29:26.990 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:26.990 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.249 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:29:27.249 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:27.249 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.508 [2024-10-01 20:28:22.505982] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:29:27.508 [2024-10-01 20:28:22.506020] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:27.508 [2024-10-01 20:28:22.506139] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:27.508 [2024-10-01 20:28:22.506234] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:27.508 [2024-10-01 20:28:22.506259] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:29:27.508 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:27.508 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:27.508 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:29:27.508 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:27.508 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.508 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:27.508 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.509 [2024-10-01 20:28:22.589978] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:29:27.509 [2024-10-01 20:28:22.590334] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:27.509 [2024-10-01 20:28:22.590421] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:29:27.509 [2024-10-01 20:28:22.590447] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:27.509 [2024-10-01 20:28:22.593604] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:27.509 pt2 00:29:27.509 [2024-10-01 20:28:22.593861] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:29:27.509 [2024-10-01 20:28:22.593977] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:29:27.509 [2024-10-01 20:28:22.594044] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:27.509 "name": "raid_bdev1", 00:29:27.509 "uuid": "8b29190a-dd4f-43da-af19-c2f886a47903", 00:29:27.509 "strip_size_kb": 64, 00:29:27.509 "state": "configuring", 00:29:27.509 "raid_level": "raid5f", 00:29:27.509 "superblock": true, 00:29:27.509 "num_base_bdevs": 3, 00:29:27.509 "num_base_bdevs_discovered": 1, 00:29:27.509 "num_base_bdevs_operational": 2, 00:29:27.509 "base_bdevs_list": [ 00:29:27.509 { 00:29:27.509 "name": null, 00:29:27.509 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:27.509 "is_configured": false, 00:29:27.509 "data_offset": 2048, 00:29:27.509 "data_size": 63488 00:29:27.509 }, 00:29:27.509 { 00:29:27.509 "name": "pt2", 00:29:27.509 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:27.509 "is_configured": true, 00:29:27.509 "data_offset": 2048, 00:29:27.509 "data_size": 63488 00:29:27.509 }, 00:29:27.509 { 00:29:27.509 "name": null, 00:29:27.509 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:27.509 "is_configured": false, 00:29:27.509 "data_offset": 2048, 00:29:27.509 "data_size": 63488 00:29:27.509 } 00:29:27.509 ] 00:29:27.509 }' 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:27.509 20:28:22 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.121 [2024-10-01 20:28:23.118299] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:29:28.121 [2024-10-01 20:28:23.118530] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:28.121 [2024-10-01 20:28:23.118609] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:29:28.121 [2024-10-01 20:28:23.118888] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:28.121 [2024-10-01 20:28:23.119502] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:28.121 [2024-10-01 20:28:23.119547] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:29:28.121 [2024-10-01 20:28:23.119660] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:29:28.121 [2024-10-01 20:28:23.119729] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:29:28.121 [2024-10-01 20:28:23.119888] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:29:28.121 [2024-10-01 20:28:23.119912] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:29:28.121 [2024-10-01 20:28:23.120205] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:29:28.121 [2024-10-01 20:28:23.125166] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:29:28.121 [2024-10-01 20:28:23.125194] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:29:28.121 pt3 00:29:28.121 [2024-10-01 20:28:23.125572] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:28.121 "name": "raid_bdev1", 00:29:28.121 "uuid": "8b29190a-dd4f-43da-af19-c2f886a47903", 00:29:28.121 "strip_size_kb": 64, 00:29:28.121 "state": "online", 00:29:28.121 "raid_level": "raid5f", 00:29:28.121 "superblock": true, 00:29:28.121 "num_base_bdevs": 3, 00:29:28.121 "num_base_bdevs_discovered": 2, 00:29:28.121 "num_base_bdevs_operational": 2, 00:29:28.121 "base_bdevs_list": [ 00:29:28.121 { 00:29:28.121 "name": null, 00:29:28.121 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:28.121 "is_configured": false, 00:29:28.121 "data_offset": 2048, 00:29:28.121 "data_size": 63488 00:29:28.121 }, 00:29:28.121 { 00:29:28.121 "name": "pt2", 00:29:28.121 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:28.121 "is_configured": true, 00:29:28.121 "data_offset": 2048, 00:29:28.121 "data_size": 63488 00:29:28.121 }, 00:29:28.121 { 00:29:28.121 "name": "pt3", 00:29:28.121 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:28.121 "is_configured": true, 00:29:28.121 "data_offset": 2048, 00:29:28.121 "data_size": 63488 00:29:28.121 } 00:29:28.121 ] 00:29:28.121 }' 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:28.121 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.689 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:29:28.689 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:28.689 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.689 [2024-10-01 20:28:23.647306] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:29:28.689 [2024-10-01 20:28:23.647491] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:28.689 [2024-10-01 20:28:23.647622] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:28.689 [2024-10-01 20:28:23.647727] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:28.689 [2024-10-01 20:28:23.647752] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:29:28.689 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:28.689 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:28.689 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:29:28.689 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.690 [2024-10-01 20:28:23.727404] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:29:28.690 [2024-10-01 20:28:23.727475] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:28.690 [2024-10-01 20:28:23.727508] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:29:28.690 [2024-10-01 20:28:23.727523] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:28.690 [2024-10-01 20:28:23.731194] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:28.690 [2024-10-01 20:28:23.731254] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:29:28.690 [2024-10-01 20:28:23.731457] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:29:28.690 [2024-10-01 20:28:23.731518] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:29:28.690 [2024-10-01 20:28:23.731857] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:29:28.690 [2024-10-01 20:28:23.731881] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:29:28.690 [2024-10-01 20:28:23.731905] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:29:28.690 [2024-10-01 20:28:23.731983] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:29:28.690 pt1 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:28.690 "name": "raid_bdev1", 00:29:28.690 "uuid": "8b29190a-dd4f-43da-af19-c2f886a47903", 00:29:28.690 "strip_size_kb": 64, 00:29:28.690 "state": "configuring", 00:29:28.690 "raid_level": "raid5f", 00:29:28.690 "superblock": true, 00:29:28.690 "num_base_bdevs": 3, 00:29:28.690 "num_base_bdevs_discovered": 1, 00:29:28.690 "num_base_bdevs_operational": 2, 00:29:28.690 "base_bdevs_list": [ 00:29:28.690 { 00:29:28.690 "name": null, 00:29:28.690 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:28.690 "is_configured": false, 00:29:28.690 "data_offset": 2048, 00:29:28.690 "data_size": 63488 00:29:28.690 }, 00:29:28.690 { 00:29:28.690 "name": "pt2", 00:29:28.690 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:28.690 "is_configured": true, 00:29:28.690 "data_offset": 2048, 00:29:28.690 "data_size": 63488 00:29:28.690 }, 00:29:28.690 { 00:29:28.690 "name": null, 00:29:28.690 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:28.690 "is_configured": false, 00:29:28.690 "data_offset": 2048, 00:29:28.690 "data_size": 63488 00:29:28.690 } 00:29:28.690 ] 00:29:28.690 }' 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:28.690 20:28:23 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:29.258 [2024-10-01 20:28:24.311731] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:29:29.258 [2024-10-01 20:28:24.311991] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:29.258 [2024-10-01 20:28:24.312078] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:29:29.258 [2024-10-01 20:28:24.312354] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:29.258 [2024-10-01 20:28:24.313006] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:29.258 [2024-10-01 20:28:24.313045] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:29:29.258 [2024-10-01 20:28:24.313159] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:29:29.258 [2024-10-01 20:28:24.313191] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:29:29.258 [2024-10-01 20:28:24.313383] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:29:29.258 [2024-10-01 20:28:24.313399] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:29:29.258 [2024-10-01 20:28:24.313761] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:29:29.258 [2024-10-01 20:28:24.318748] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:29:29.258 [2024-10-01 20:28:24.318780] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:29:29.258 pt3 00:29:29.258 [2024-10-01 20:28:24.319084] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:29.258 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:29.259 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:29.259 "name": "raid_bdev1", 00:29:29.259 "uuid": "8b29190a-dd4f-43da-af19-c2f886a47903", 00:29:29.259 "strip_size_kb": 64, 00:29:29.259 "state": "online", 00:29:29.259 "raid_level": "raid5f", 00:29:29.259 "superblock": true, 00:29:29.259 "num_base_bdevs": 3, 00:29:29.259 "num_base_bdevs_discovered": 2, 00:29:29.259 "num_base_bdevs_operational": 2, 00:29:29.259 "base_bdevs_list": [ 00:29:29.259 { 00:29:29.259 "name": null, 00:29:29.259 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:29.259 "is_configured": false, 00:29:29.259 "data_offset": 2048, 00:29:29.259 "data_size": 63488 00:29:29.259 }, 00:29:29.259 { 00:29:29.259 "name": "pt2", 00:29:29.259 "uuid": "00000000-0000-0000-0000-000000000002", 00:29:29.259 "is_configured": true, 00:29:29.259 "data_offset": 2048, 00:29:29.259 "data_size": 63488 00:29:29.259 }, 00:29:29.259 { 00:29:29.259 "name": "pt3", 00:29:29.259 "uuid": "00000000-0000-0000-0000-000000000003", 00:29:29.259 "is_configured": true, 00:29:29.259 "data_offset": 2048, 00:29:29.259 "data_size": 63488 00:29:29.259 } 00:29:29.259 ] 00:29:29.259 }' 00:29:29.259 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:29.259 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:29:29.826 [2024-10-01 20:28:24.869111] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 8b29190a-dd4f-43da-af19-c2f886a47903 '!=' 8b29190a-dd4f-43da-af19-c2f886a47903 ']' 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 82375 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 82375 ']' 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # kill -0 82375 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@955 -- # uname 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 82375 00:29:29.826 killing process with pid 82375 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 82375' 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@969 -- # kill 82375 00:29:29.826 [2024-10-01 20:28:24.950535] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:29:29.826 20:28:24 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@974 -- # wait 82375 00:29:29.826 [2024-10-01 20:28:24.950658] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:29.826 [2024-10-01 20:28:24.950762] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:29.826 [2024-10-01 20:28:24.950787] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:29:30.085 [2024-10-01 20:28:25.250619] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:29:32.618 20:28:27 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:29:32.618 00:29:32.618 real 0m9.942s 00:29:32.618 user 0m15.343s 00:29:32.618 sys 0m1.433s 00:29:32.618 ************************************ 00:29:32.618 END TEST raid5f_superblock_test 00:29:32.618 ************************************ 00:29:32.618 20:28:27 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:29:32.618 20:28:27 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:29:32.618 20:28:27 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:29:32.618 20:28:27 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 3 false false true 00:29:32.618 20:28:27 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:29:32.618 20:28:27 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:29:32.618 20:28:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:29:32.618 ************************************ 00:29:32.618 START TEST raid5f_rebuild_test 00:29:32.618 ************************************ 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid5f 3 false false true 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=82836 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 82836 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@831 -- # '[' -z 82836 ']' 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:29:32.618 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:29:32.618 20:28:27 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:32.618 [2024-10-01 20:28:27.554877] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:29:32.618 I/O size of 3145728 is greater than zero copy threshold (65536). 00:29:32.618 Zero copy mechanism will not be used. 00:29:32.618 [2024-10-01 20:28:27.555372] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid82836 ] 00:29:32.618 [2024-10-01 20:28:27.734567] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:32.877 [2024-10-01 20:28:28.029031] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:29:33.136 [2024-10-01 20:28:28.259730] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:33.136 [2024-10-01 20:28:28.259816] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # return 0 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:33.703 BaseBdev1_malloc 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:33.703 [2024-10-01 20:28:28.776320] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:29:33.703 [2024-10-01 20:28:28.776407] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:33.703 [2024-10-01 20:28:28.776441] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:29:33.703 [2024-10-01 20:28:28.776466] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:33.703 [2024-10-01 20:28:28.779921] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:33.703 [2024-10-01 20:28:28.779982] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:29:33.703 BaseBdev1 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:33.703 BaseBdev2_malloc 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:33.703 [2024-10-01 20:28:28.835096] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:29:33.703 [2024-10-01 20:28:28.835192] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:33.703 [2024-10-01 20:28:28.835240] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:29:33.703 [2024-10-01 20:28:28.835260] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:33.703 [2024-10-01 20:28:28.838451] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:33.703 [2024-10-01 20:28:28.838670] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:29:33.703 BaseBdev2 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:33.703 BaseBdev3_malloc 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:33.703 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:29:33.704 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:33.704 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:33.704 [2024-10-01 20:28:28.893550] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:29:33.704 [2024-10-01 20:28:28.893642] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:33.704 [2024-10-01 20:28:28.893678] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:29:33.704 [2024-10-01 20:28:28.893698] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:33.704 [2024-10-01 20:28:28.897222] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:33.704 [2024-10-01 20:28:28.897295] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:29:33.704 BaseBdev3 00:29:33.704 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:33.704 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:29:33.704 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:33.704 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:33.704 spare_malloc 00:29:33.704 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:33.704 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:29:33.704 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:33.704 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:33.964 spare_delay 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:33.964 [2024-10-01 20:28:28.962209] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:29:33.964 [2024-10-01 20:28:28.962450] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:33.964 [2024-10-01 20:28:28.962491] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:29:33.964 [2024-10-01 20:28:28.962513] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:33.964 [2024-10-01 20:28:28.965741] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:33.964 [2024-10-01 20:28:28.965796] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:29:33.964 spare 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:33.964 [2024-10-01 20:28:28.970212] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:33.964 [2024-10-01 20:28:28.972933] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:33.964 [2024-10-01 20:28:28.973031] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:33.964 [2024-10-01 20:28:28.973162] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:29:33.964 [2024-10-01 20:28:28.973179] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:29:33.964 [2024-10-01 20:28:28.973553] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:29:33.964 [2024-10-01 20:28:28.979808] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:29:33.964 [2024-10-01 20:28:28.979841] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:29:33.964 [2024-10-01 20:28:28.980203] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:33.964 20:28:28 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:33.964 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:33.964 "name": "raid_bdev1", 00:29:33.964 "uuid": "2e849680-af2f-460a-a5ec-8e32f33a1e9f", 00:29:33.964 "strip_size_kb": 64, 00:29:33.964 "state": "online", 00:29:33.964 "raid_level": "raid5f", 00:29:33.964 "superblock": false, 00:29:33.964 "num_base_bdevs": 3, 00:29:33.964 "num_base_bdevs_discovered": 3, 00:29:33.964 "num_base_bdevs_operational": 3, 00:29:33.964 "base_bdevs_list": [ 00:29:33.964 { 00:29:33.964 "name": "BaseBdev1", 00:29:33.964 "uuid": "f757bccb-b76c-5a94-8307-b6fb636d050d", 00:29:33.964 "is_configured": true, 00:29:33.964 "data_offset": 0, 00:29:33.964 "data_size": 65536 00:29:33.964 }, 00:29:33.964 { 00:29:33.964 "name": "BaseBdev2", 00:29:33.964 "uuid": "8645ef30-08b7-5b67-a749-3b4a0459b1f8", 00:29:33.964 "is_configured": true, 00:29:33.964 "data_offset": 0, 00:29:33.964 "data_size": 65536 00:29:33.964 }, 00:29:33.964 { 00:29:33.964 "name": "BaseBdev3", 00:29:33.964 "uuid": "68d9b497-6f69-5c27-9859-742312b9e150", 00:29:33.964 "is_configured": true, 00:29:33.964 "data_offset": 0, 00:29:33.964 "data_size": 65536 00:29:33.964 } 00:29:33.964 ] 00:29:33.964 }' 00:29:33.964 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:33.964 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:29:34.531 [2024-10-01 20:28:29.494752] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=131072 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:29:34.531 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:29:34.789 [2024-10-01 20:28:29.902766] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:29:34.789 /dev/nbd0 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:29:34.789 1+0 records in 00:29:34.789 1+0 records out 00:29:34.789 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000455554 s, 9.0 MB/s 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 128 00:29:34.789 20:28:29 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=512 oflag=direct 00:29:35.355 512+0 records in 00:29:35.356 512+0 records out 00:29:35.356 67108864 bytes (67 MB, 64 MiB) copied, 0.502082 s, 134 MB/s 00:29:35.356 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:29:35.356 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:29:35.356 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:29:35.356 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:29:35.356 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:29:35.356 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:29:35.356 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:29:35.614 [2024-10-01 20:28:30.853436] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:35.614 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:29:35.614 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:29:35.614 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:29:35.614 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:29:35.614 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:29:35.614 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:29:35.614 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:29:35.614 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:29:35.614 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:29:35.614 20:28:30 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:35.614 20:28:30 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:35.873 [2024-10-01 20:28:30.871278] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:35.873 "name": "raid_bdev1", 00:29:35.873 "uuid": "2e849680-af2f-460a-a5ec-8e32f33a1e9f", 00:29:35.873 "strip_size_kb": 64, 00:29:35.873 "state": "online", 00:29:35.873 "raid_level": "raid5f", 00:29:35.873 "superblock": false, 00:29:35.873 "num_base_bdevs": 3, 00:29:35.873 "num_base_bdevs_discovered": 2, 00:29:35.873 "num_base_bdevs_operational": 2, 00:29:35.873 "base_bdevs_list": [ 00:29:35.873 { 00:29:35.873 "name": null, 00:29:35.873 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:35.873 "is_configured": false, 00:29:35.873 "data_offset": 0, 00:29:35.873 "data_size": 65536 00:29:35.873 }, 00:29:35.873 { 00:29:35.873 "name": "BaseBdev2", 00:29:35.873 "uuid": "8645ef30-08b7-5b67-a749-3b4a0459b1f8", 00:29:35.873 "is_configured": true, 00:29:35.873 "data_offset": 0, 00:29:35.873 "data_size": 65536 00:29:35.873 }, 00:29:35.873 { 00:29:35.873 "name": "BaseBdev3", 00:29:35.873 "uuid": "68d9b497-6f69-5c27-9859-742312b9e150", 00:29:35.873 "is_configured": true, 00:29:35.873 "data_offset": 0, 00:29:35.873 "data_size": 65536 00:29:35.873 } 00:29:35.873 ] 00:29:35.873 }' 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:35.873 20:28:30 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:36.442 20:28:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:29:36.442 20:28:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:36.442 20:28:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:36.442 [2024-10-01 20:28:31.403461] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:29:36.442 [2024-10-01 20:28:31.418300] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b680 00:29:36.442 20:28:31 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:36.442 20:28:31 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:29:36.442 [2024-10-01 20:28:31.426055] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:29:37.379 "name": "raid_bdev1", 00:29:37.379 "uuid": "2e849680-af2f-460a-a5ec-8e32f33a1e9f", 00:29:37.379 "strip_size_kb": 64, 00:29:37.379 "state": "online", 00:29:37.379 "raid_level": "raid5f", 00:29:37.379 "superblock": false, 00:29:37.379 "num_base_bdevs": 3, 00:29:37.379 "num_base_bdevs_discovered": 3, 00:29:37.379 "num_base_bdevs_operational": 3, 00:29:37.379 "process": { 00:29:37.379 "type": "rebuild", 00:29:37.379 "target": "spare", 00:29:37.379 "progress": { 00:29:37.379 "blocks": 18432, 00:29:37.379 "percent": 14 00:29:37.379 } 00:29:37.379 }, 00:29:37.379 "base_bdevs_list": [ 00:29:37.379 { 00:29:37.379 "name": "spare", 00:29:37.379 "uuid": "80e66d6b-af23-5ccc-b277-cef4ce94f1f9", 00:29:37.379 "is_configured": true, 00:29:37.379 "data_offset": 0, 00:29:37.379 "data_size": 65536 00:29:37.379 }, 00:29:37.379 { 00:29:37.379 "name": "BaseBdev2", 00:29:37.379 "uuid": "8645ef30-08b7-5b67-a749-3b4a0459b1f8", 00:29:37.379 "is_configured": true, 00:29:37.379 "data_offset": 0, 00:29:37.379 "data_size": 65536 00:29:37.379 }, 00:29:37.379 { 00:29:37.379 "name": "BaseBdev3", 00:29:37.379 "uuid": "68d9b497-6f69-5c27-9859-742312b9e150", 00:29:37.379 "is_configured": true, 00:29:37.379 "data_offset": 0, 00:29:37.379 "data_size": 65536 00:29:37.379 } 00:29:37.379 ] 00:29:37.379 }' 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:37.379 20:28:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:37.379 [2024-10-01 20:28:32.595648] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:29:37.638 [2024-10-01 20:28:32.641098] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:29:37.638 [2024-10-01 20:28:32.641227] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:37.638 [2024-10-01 20:28:32.641262] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:29:37.638 [2024-10-01 20:28:32.641273] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:37.638 "name": "raid_bdev1", 00:29:37.638 "uuid": "2e849680-af2f-460a-a5ec-8e32f33a1e9f", 00:29:37.638 "strip_size_kb": 64, 00:29:37.638 "state": "online", 00:29:37.638 "raid_level": "raid5f", 00:29:37.638 "superblock": false, 00:29:37.638 "num_base_bdevs": 3, 00:29:37.638 "num_base_bdevs_discovered": 2, 00:29:37.638 "num_base_bdevs_operational": 2, 00:29:37.638 "base_bdevs_list": [ 00:29:37.638 { 00:29:37.638 "name": null, 00:29:37.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:37.638 "is_configured": false, 00:29:37.638 "data_offset": 0, 00:29:37.638 "data_size": 65536 00:29:37.638 }, 00:29:37.638 { 00:29:37.638 "name": "BaseBdev2", 00:29:37.638 "uuid": "8645ef30-08b7-5b67-a749-3b4a0459b1f8", 00:29:37.638 "is_configured": true, 00:29:37.638 "data_offset": 0, 00:29:37.638 "data_size": 65536 00:29:37.638 }, 00:29:37.638 { 00:29:37.638 "name": "BaseBdev3", 00:29:37.638 "uuid": "68d9b497-6f69-5c27-9859-742312b9e150", 00:29:37.638 "is_configured": true, 00:29:37.638 "data_offset": 0, 00:29:37.638 "data_size": 65536 00:29:37.638 } 00:29:37.638 ] 00:29:37.638 }' 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:37.638 20:28:32 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:29:38.206 "name": "raid_bdev1", 00:29:38.206 "uuid": "2e849680-af2f-460a-a5ec-8e32f33a1e9f", 00:29:38.206 "strip_size_kb": 64, 00:29:38.206 "state": "online", 00:29:38.206 "raid_level": "raid5f", 00:29:38.206 "superblock": false, 00:29:38.206 "num_base_bdevs": 3, 00:29:38.206 "num_base_bdevs_discovered": 2, 00:29:38.206 "num_base_bdevs_operational": 2, 00:29:38.206 "base_bdevs_list": [ 00:29:38.206 { 00:29:38.206 "name": null, 00:29:38.206 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:38.206 "is_configured": false, 00:29:38.206 "data_offset": 0, 00:29:38.206 "data_size": 65536 00:29:38.206 }, 00:29:38.206 { 00:29:38.206 "name": "BaseBdev2", 00:29:38.206 "uuid": "8645ef30-08b7-5b67-a749-3b4a0459b1f8", 00:29:38.206 "is_configured": true, 00:29:38.206 "data_offset": 0, 00:29:38.206 "data_size": 65536 00:29:38.206 }, 00:29:38.206 { 00:29:38.206 "name": "BaseBdev3", 00:29:38.206 "uuid": "68d9b497-6f69-5c27-9859-742312b9e150", 00:29:38.206 "is_configured": true, 00:29:38.206 "data_offset": 0, 00:29:38.206 "data_size": 65536 00:29:38.206 } 00:29:38.206 ] 00:29:38.206 }' 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:38.206 [2024-10-01 20:28:33.351221] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:29:38.206 [2024-10-01 20:28:33.365396] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b750 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:38.206 20:28:33 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:29:38.206 [2024-10-01 20:28:33.373473] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:29:39.181 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:39.181 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:29:39.181 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:29:39.181 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:29:39.181 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:29:39.181 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:39.181 20:28:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:39.181 20:28:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:39.181 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:39.181 20:28:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:39.448 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:29:39.448 "name": "raid_bdev1", 00:29:39.448 "uuid": "2e849680-af2f-460a-a5ec-8e32f33a1e9f", 00:29:39.448 "strip_size_kb": 64, 00:29:39.448 "state": "online", 00:29:39.448 "raid_level": "raid5f", 00:29:39.448 "superblock": false, 00:29:39.448 "num_base_bdevs": 3, 00:29:39.448 "num_base_bdevs_discovered": 3, 00:29:39.448 "num_base_bdevs_operational": 3, 00:29:39.449 "process": { 00:29:39.449 "type": "rebuild", 00:29:39.449 "target": "spare", 00:29:39.449 "progress": { 00:29:39.449 "blocks": 18432, 00:29:39.449 "percent": 14 00:29:39.449 } 00:29:39.449 }, 00:29:39.449 "base_bdevs_list": [ 00:29:39.449 { 00:29:39.449 "name": "spare", 00:29:39.449 "uuid": "80e66d6b-af23-5ccc-b277-cef4ce94f1f9", 00:29:39.449 "is_configured": true, 00:29:39.449 "data_offset": 0, 00:29:39.449 "data_size": 65536 00:29:39.449 }, 00:29:39.449 { 00:29:39.449 "name": "BaseBdev2", 00:29:39.449 "uuid": "8645ef30-08b7-5b67-a749-3b4a0459b1f8", 00:29:39.449 "is_configured": true, 00:29:39.449 "data_offset": 0, 00:29:39.449 "data_size": 65536 00:29:39.449 }, 00:29:39.449 { 00:29:39.449 "name": "BaseBdev3", 00:29:39.449 "uuid": "68d9b497-6f69-5c27-9859-742312b9e150", 00:29:39.449 "is_configured": true, 00:29:39.449 "data_offset": 0, 00:29:39.449 "data_size": 65536 00:29:39.449 } 00:29:39.449 ] 00:29:39.449 }' 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=653 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:29:39.449 "name": "raid_bdev1", 00:29:39.449 "uuid": "2e849680-af2f-460a-a5ec-8e32f33a1e9f", 00:29:39.449 "strip_size_kb": 64, 00:29:39.449 "state": "online", 00:29:39.449 "raid_level": "raid5f", 00:29:39.449 "superblock": false, 00:29:39.449 "num_base_bdevs": 3, 00:29:39.449 "num_base_bdevs_discovered": 3, 00:29:39.449 "num_base_bdevs_operational": 3, 00:29:39.449 "process": { 00:29:39.449 "type": "rebuild", 00:29:39.449 "target": "spare", 00:29:39.449 "progress": { 00:29:39.449 "blocks": 22528, 00:29:39.449 "percent": 17 00:29:39.449 } 00:29:39.449 }, 00:29:39.449 "base_bdevs_list": [ 00:29:39.449 { 00:29:39.449 "name": "spare", 00:29:39.449 "uuid": "80e66d6b-af23-5ccc-b277-cef4ce94f1f9", 00:29:39.449 "is_configured": true, 00:29:39.449 "data_offset": 0, 00:29:39.449 "data_size": 65536 00:29:39.449 }, 00:29:39.449 { 00:29:39.449 "name": "BaseBdev2", 00:29:39.449 "uuid": "8645ef30-08b7-5b67-a749-3b4a0459b1f8", 00:29:39.449 "is_configured": true, 00:29:39.449 "data_offset": 0, 00:29:39.449 "data_size": 65536 00:29:39.449 }, 00:29:39.449 { 00:29:39.449 "name": "BaseBdev3", 00:29:39.449 "uuid": "68d9b497-6f69-5c27-9859-742312b9e150", 00:29:39.449 "is_configured": true, 00:29:39.449 "data_offset": 0, 00:29:39.449 "data_size": 65536 00:29:39.449 } 00:29:39.449 ] 00:29:39.449 }' 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:39.449 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:29:39.708 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:29:39.708 20:28:34 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:29:40.645 "name": "raid_bdev1", 00:29:40.645 "uuid": "2e849680-af2f-460a-a5ec-8e32f33a1e9f", 00:29:40.645 "strip_size_kb": 64, 00:29:40.645 "state": "online", 00:29:40.645 "raid_level": "raid5f", 00:29:40.645 "superblock": false, 00:29:40.645 "num_base_bdevs": 3, 00:29:40.645 "num_base_bdevs_discovered": 3, 00:29:40.645 "num_base_bdevs_operational": 3, 00:29:40.645 "process": { 00:29:40.645 "type": "rebuild", 00:29:40.645 "target": "spare", 00:29:40.645 "progress": { 00:29:40.645 "blocks": 47104, 00:29:40.645 "percent": 35 00:29:40.645 } 00:29:40.645 }, 00:29:40.645 "base_bdevs_list": [ 00:29:40.645 { 00:29:40.645 "name": "spare", 00:29:40.645 "uuid": "80e66d6b-af23-5ccc-b277-cef4ce94f1f9", 00:29:40.645 "is_configured": true, 00:29:40.645 "data_offset": 0, 00:29:40.645 "data_size": 65536 00:29:40.645 }, 00:29:40.645 { 00:29:40.645 "name": "BaseBdev2", 00:29:40.645 "uuid": "8645ef30-08b7-5b67-a749-3b4a0459b1f8", 00:29:40.645 "is_configured": true, 00:29:40.645 "data_offset": 0, 00:29:40.645 "data_size": 65536 00:29:40.645 }, 00:29:40.645 { 00:29:40.645 "name": "BaseBdev3", 00:29:40.645 "uuid": "68d9b497-6f69-5c27-9859-742312b9e150", 00:29:40.645 "is_configured": true, 00:29:40.645 "data_offset": 0, 00:29:40.645 "data_size": 65536 00:29:40.645 } 00:29:40.645 ] 00:29:40.645 }' 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:29:40.645 20:28:35 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:29:42.023 20:28:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:29:42.023 20:28:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:42.023 20:28:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:29:42.023 20:28:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:29:42.023 20:28:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:29:42.023 20:28:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:29:42.023 20:28:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:42.023 20:28:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:42.023 20:28:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:42.023 20:28:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:42.023 20:28:36 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:42.023 20:28:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:29:42.023 "name": "raid_bdev1", 00:29:42.023 "uuid": "2e849680-af2f-460a-a5ec-8e32f33a1e9f", 00:29:42.023 "strip_size_kb": 64, 00:29:42.023 "state": "online", 00:29:42.023 "raid_level": "raid5f", 00:29:42.023 "superblock": false, 00:29:42.023 "num_base_bdevs": 3, 00:29:42.023 "num_base_bdevs_discovered": 3, 00:29:42.023 "num_base_bdevs_operational": 3, 00:29:42.023 "process": { 00:29:42.023 "type": "rebuild", 00:29:42.023 "target": "spare", 00:29:42.023 "progress": { 00:29:42.023 "blocks": 69632, 00:29:42.023 "percent": 53 00:29:42.023 } 00:29:42.023 }, 00:29:42.023 "base_bdevs_list": [ 00:29:42.023 { 00:29:42.023 "name": "spare", 00:29:42.023 "uuid": "80e66d6b-af23-5ccc-b277-cef4ce94f1f9", 00:29:42.023 "is_configured": true, 00:29:42.023 "data_offset": 0, 00:29:42.023 "data_size": 65536 00:29:42.023 }, 00:29:42.023 { 00:29:42.023 "name": "BaseBdev2", 00:29:42.023 "uuid": "8645ef30-08b7-5b67-a749-3b4a0459b1f8", 00:29:42.023 "is_configured": true, 00:29:42.023 "data_offset": 0, 00:29:42.023 "data_size": 65536 00:29:42.023 }, 00:29:42.023 { 00:29:42.023 "name": "BaseBdev3", 00:29:42.023 "uuid": "68d9b497-6f69-5c27-9859-742312b9e150", 00:29:42.023 "is_configured": true, 00:29:42.023 "data_offset": 0, 00:29:42.023 "data_size": 65536 00:29:42.023 } 00:29:42.023 ] 00:29:42.023 }' 00:29:42.023 20:28:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:29:42.023 20:28:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:42.023 20:28:36 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:29:42.023 20:28:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:29:42.023 20:28:37 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:29:42.962 20:28:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:29:42.962 20:28:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:42.962 20:28:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:29:42.962 20:28:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:29:42.962 20:28:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:29:42.962 20:28:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:29:42.962 20:28:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:42.962 20:28:38 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:42.962 20:28:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:42.962 20:28:38 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:42.962 20:28:38 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:42.962 20:28:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:29:42.962 "name": "raid_bdev1", 00:29:42.962 "uuid": "2e849680-af2f-460a-a5ec-8e32f33a1e9f", 00:29:42.962 "strip_size_kb": 64, 00:29:42.962 "state": "online", 00:29:42.962 "raid_level": "raid5f", 00:29:42.962 "superblock": false, 00:29:42.962 "num_base_bdevs": 3, 00:29:42.962 "num_base_bdevs_discovered": 3, 00:29:42.962 "num_base_bdevs_operational": 3, 00:29:42.962 "process": { 00:29:42.962 "type": "rebuild", 00:29:42.962 "target": "spare", 00:29:42.962 "progress": { 00:29:42.962 "blocks": 94208, 00:29:42.962 "percent": 71 00:29:42.962 } 00:29:42.962 }, 00:29:42.962 "base_bdevs_list": [ 00:29:42.962 { 00:29:42.962 "name": "spare", 00:29:42.962 "uuid": "80e66d6b-af23-5ccc-b277-cef4ce94f1f9", 00:29:42.962 "is_configured": true, 00:29:42.962 "data_offset": 0, 00:29:42.962 "data_size": 65536 00:29:42.962 }, 00:29:42.962 { 00:29:42.962 "name": "BaseBdev2", 00:29:42.962 "uuid": "8645ef30-08b7-5b67-a749-3b4a0459b1f8", 00:29:42.962 "is_configured": true, 00:29:42.962 "data_offset": 0, 00:29:42.962 "data_size": 65536 00:29:42.962 }, 00:29:42.962 { 00:29:42.962 "name": "BaseBdev3", 00:29:42.962 "uuid": "68d9b497-6f69-5c27-9859-742312b9e150", 00:29:42.962 "is_configured": true, 00:29:42.962 "data_offset": 0, 00:29:42.962 "data_size": 65536 00:29:42.962 } 00:29:42.962 ] 00:29:42.962 }' 00:29:42.963 20:28:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:29:42.963 20:28:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:42.963 20:28:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:29:42.963 20:28:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:29:42.963 20:28:38 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:29:44.342 "name": "raid_bdev1", 00:29:44.342 "uuid": "2e849680-af2f-460a-a5ec-8e32f33a1e9f", 00:29:44.342 "strip_size_kb": 64, 00:29:44.342 "state": "online", 00:29:44.342 "raid_level": "raid5f", 00:29:44.342 "superblock": false, 00:29:44.342 "num_base_bdevs": 3, 00:29:44.342 "num_base_bdevs_discovered": 3, 00:29:44.342 "num_base_bdevs_operational": 3, 00:29:44.342 "process": { 00:29:44.342 "type": "rebuild", 00:29:44.342 "target": "spare", 00:29:44.342 "progress": { 00:29:44.342 "blocks": 116736, 00:29:44.342 "percent": 89 00:29:44.342 } 00:29:44.342 }, 00:29:44.342 "base_bdevs_list": [ 00:29:44.342 { 00:29:44.342 "name": "spare", 00:29:44.342 "uuid": "80e66d6b-af23-5ccc-b277-cef4ce94f1f9", 00:29:44.342 "is_configured": true, 00:29:44.342 "data_offset": 0, 00:29:44.342 "data_size": 65536 00:29:44.342 }, 00:29:44.342 { 00:29:44.342 "name": "BaseBdev2", 00:29:44.342 "uuid": "8645ef30-08b7-5b67-a749-3b4a0459b1f8", 00:29:44.342 "is_configured": true, 00:29:44.342 "data_offset": 0, 00:29:44.342 "data_size": 65536 00:29:44.342 }, 00:29:44.342 { 00:29:44.342 "name": "BaseBdev3", 00:29:44.342 "uuid": "68d9b497-6f69-5c27-9859-742312b9e150", 00:29:44.342 "is_configured": true, 00:29:44.342 "data_offset": 0, 00:29:44.342 "data_size": 65536 00:29:44.342 } 00:29:44.342 ] 00:29:44.342 }' 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:29:44.342 20:28:39 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:29:44.910 [2024-10-01 20:28:39.861949] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:29:44.910 [2024-10-01 20:28:39.862110] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:29:44.910 [2024-10-01 20:28:39.862181] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:45.169 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:29:45.169 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:45.169 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:29:45.169 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:29:45.169 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:29:45.169 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:29:45.169 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:45.169 20:28:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:45.169 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:45.169 20:28:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:45.169 20:28:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:45.169 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:29:45.169 "name": "raid_bdev1", 00:29:45.169 "uuid": "2e849680-af2f-460a-a5ec-8e32f33a1e9f", 00:29:45.169 "strip_size_kb": 64, 00:29:45.169 "state": "online", 00:29:45.169 "raid_level": "raid5f", 00:29:45.169 "superblock": false, 00:29:45.169 "num_base_bdevs": 3, 00:29:45.169 "num_base_bdevs_discovered": 3, 00:29:45.169 "num_base_bdevs_operational": 3, 00:29:45.169 "base_bdevs_list": [ 00:29:45.169 { 00:29:45.169 "name": "spare", 00:29:45.169 "uuid": "80e66d6b-af23-5ccc-b277-cef4ce94f1f9", 00:29:45.169 "is_configured": true, 00:29:45.169 "data_offset": 0, 00:29:45.169 "data_size": 65536 00:29:45.169 }, 00:29:45.169 { 00:29:45.169 "name": "BaseBdev2", 00:29:45.169 "uuid": "8645ef30-08b7-5b67-a749-3b4a0459b1f8", 00:29:45.169 "is_configured": true, 00:29:45.169 "data_offset": 0, 00:29:45.169 "data_size": 65536 00:29:45.169 }, 00:29:45.169 { 00:29:45.169 "name": "BaseBdev3", 00:29:45.169 "uuid": "68d9b497-6f69-5c27-9859-742312b9e150", 00:29:45.169 "is_configured": true, 00:29:45.169 "data_offset": 0, 00:29:45.169 "data_size": 65536 00:29:45.169 } 00:29:45.169 ] 00:29:45.169 }' 00:29:45.169 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:29:45.455 "name": "raid_bdev1", 00:29:45.455 "uuid": "2e849680-af2f-460a-a5ec-8e32f33a1e9f", 00:29:45.455 "strip_size_kb": 64, 00:29:45.455 "state": "online", 00:29:45.455 "raid_level": "raid5f", 00:29:45.455 "superblock": false, 00:29:45.455 "num_base_bdevs": 3, 00:29:45.455 "num_base_bdevs_discovered": 3, 00:29:45.455 "num_base_bdevs_operational": 3, 00:29:45.455 "base_bdevs_list": [ 00:29:45.455 { 00:29:45.455 "name": "spare", 00:29:45.455 "uuid": "80e66d6b-af23-5ccc-b277-cef4ce94f1f9", 00:29:45.455 "is_configured": true, 00:29:45.455 "data_offset": 0, 00:29:45.455 "data_size": 65536 00:29:45.455 }, 00:29:45.455 { 00:29:45.455 "name": "BaseBdev2", 00:29:45.455 "uuid": "8645ef30-08b7-5b67-a749-3b4a0459b1f8", 00:29:45.455 "is_configured": true, 00:29:45.455 "data_offset": 0, 00:29:45.455 "data_size": 65536 00:29:45.455 }, 00:29:45.455 { 00:29:45.455 "name": "BaseBdev3", 00:29:45.455 "uuid": "68d9b497-6f69-5c27-9859-742312b9e150", 00:29:45.455 "is_configured": true, 00:29:45.455 "data_offset": 0, 00:29:45.455 "data_size": 65536 00:29:45.455 } 00:29:45.455 ] 00:29:45.455 }' 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:45.455 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:45.456 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:45.456 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:45.456 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:45.456 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:45.456 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:45.714 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:45.714 20:28:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:45.714 20:28:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:45.714 20:28:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:45.714 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:45.714 "name": "raid_bdev1", 00:29:45.714 "uuid": "2e849680-af2f-460a-a5ec-8e32f33a1e9f", 00:29:45.714 "strip_size_kb": 64, 00:29:45.714 "state": "online", 00:29:45.714 "raid_level": "raid5f", 00:29:45.714 "superblock": false, 00:29:45.714 "num_base_bdevs": 3, 00:29:45.714 "num_base_bdevs_discovered": 3, 00:29:45.714 "num_base_bdevs_operational": 3, 00:29:45.714 "base_bdevs_list": [ 00:29:45.714 { 00:29:45.714 "name": "spare", 00:29:45.714 "uuid": "80e66d6b-af23-5ccc-b277-cef4ce94f1f9", 00:29:45.714 "is_configured": true, 00:29:45.714 "data_offset": 0, 00:29:45.714 "data_size": 65536 00:29:45.714 }, 00:29:45.714 { 00:29:45.714 "name": "BaseBdev2", 00:29:45.714 "uuid": "8645ef30-08b7-5b67-a749-3b4a0459b1f8", 00:29:45.714 "is_configured": true, 00:29:45.714 "data_offset": 0, 00:29:45.714 "data_size": 65536 00:29:45.714 }, 00:29:45.714 { 00:29:45.714 "name": "BaseBdev3", 00:29:45.714 "uuid": "68d9b497-6f69-5c27-9859-742312b9e150", 00:29:45.714 "is_configured": true, 00:29:45.714 "data_offset": 0, 00:29:45.714 "data_size": 65536 00:29:45.714 } 00:29:45.714 ] 00:29:45.714 }' 00:29:45.714 20:28:40 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:45.714 20:28:40 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:46.281 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:29:46.281 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:46.281 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:46.281 [2024-10-01 20:28:41.232183] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:29:46.281 [2024-10-01 20:28:41.232226] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:29:46.281 [2024-10-01 20:28:41.232341] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:29:46.281 [2024-10-01 20:28:41.232463] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:29:46.281 [2024-10-01 20:28:41.232489] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:29:46.282 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:29:46.540 /dev/nbd0 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:29:46.540 1+0 records in 00:29:46.540 1+0 records out 00:29:46.540 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000277198 s, 14.8 MB/s 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:29:46.540 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:29:46.799 /dev/nbd1 00:29:46.799 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:29:46.799 20:28:41 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:29:46.799 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:29:46.799 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:29:46.799 20:28:41 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:29:46.799 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:29:46.799 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:29:46.799 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:29:46.799 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:29:46.799 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:29:46.799 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:29:46.799 1+0 records in 00:29:46.799 1+0 records out 00:29:46.799 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000432287 s, 9.5 MB/s 00:29:46.799 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:29:46.799 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:29:46.799 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:29:46.799 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:29:46.799 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:29:46.799 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:29:46.799 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:29:46.799 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:29:47.057 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:29:47.057 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:29:47.057 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:29:47.057 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:29:47.057 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:29:47.057 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:29:47.057 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:29:47.316 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:29:47.316 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:29:47.316 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:29:47.316 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:29:47.317 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:29:47.317 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:29:47.317 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:29:47.317 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:29:47.317 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:29:47.317 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 82836 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@950 -- # '[' -z 82836 ']' 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # kill -0 82836 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@955 -- # uname 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 82836 00:29:47.885 killing process with pid 82836 00:29:47.885 Received shutdown signal, test time was about 60.000000 seconds 00:29:47.885 00:29:47.885 Latency(us) 00:29:47.885 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:47.885 =================================================================================================================== 00:29:47.885 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 82836' 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@969 -- # kill 82836 00:29:47.885 [2024-10-01 20:28:42.884047] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:29:47.885 20:28:42 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@974 -- # wait 82836 00:29:48.146 [2024-10-01 20:28:43.266876] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:29:50.053 20:28:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:29:50.053 00:29:50.053 real 0m17.821s 00:29:50.053 user 0m22.370s 00:29:50.053 sys 0m2.305s 00:29:50.053 ************************************ 00:29:50.053 END TEST raid5f_rebuild_test 00:29:50.053 ************************************ 00:29:50.053 20:28:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:29:50.053 20:28:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:29:50.053 20:28:45 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 3 true false true 00:29:50.053 20:28:45 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:29:50.053 20:28:45 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:29:50.053 20:28:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:29:50.315 ************************************ 00:29:50.315 START TEST raid5f_rebuild_test_sb 00:29:50.315 ************************************ 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid5f 3 true false true 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=83296 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 83296 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@831 -- # '[' -z 83296 ']' 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:29:50.315 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:29:50.315 20:28:45 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:50.315 I/O size of 3145728 is greater than zero copy threshold (65536). 00:29:50.315 Zero copy mechanism will not be used. 00:29:50.315 [2024-10-01 20:28:45.446642] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:29:50.315 [2024-10-01 20:28:45.446896] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83296 ] 00:29:50.574 [2024-10-01 20:28:45.628239] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:50.833 [2024-10-01 20:28:45.903601] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:29:51.092 [2024-10-01 20:28:46.120228] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:51.092 [2024-10-01 20:28:46.120288] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:29:51.352 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:29:51.352 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # return 0 00:29:51.352 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:29:51.352 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:29:51.352 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:51.352 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:51.611 BaseBdev1_malloc 00:29:51.611 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:51.612 [2024-10-01 20:28:46.619554] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:29:51.612 [2024-10-01 20:28:46.619692] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:51.612 [2024-10-01 20:28:46.619755] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:29:51.612 [2024-10-01 20:28:46.619808] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:51.612 [2024-10-01 20:28:46.623197] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:51.612 [2024-10-01 20:28:46.623255] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:29:51.612 BaseBdev1 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:51.612 BaseBdev2_malloc 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:51.612 [2024-10-01 20:28:46.680864] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:29:51.612 [2024-10-01 20:28:46.680956] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:51.612 [2024-10-01 20:28:46.681003] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:29:51.612 [2024-10-01 20:28:46.681028] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:51.612 [2024-10-01 20:28:46.684236] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:51.612 [2024-10-01 20:28:46.684289] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:29:51.612 BaseBdev2 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:51.612 BaseBdev3_malloc 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:51.612 [2024-10-01 20:28:46.744428] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:29:51.612 [2024-10-01 20:28:46.744530] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:51.612 [2024-10-01 20:28:46.744567] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:29:51.612 [2024-10-01 20:28:46.744592] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:51.612 [2024-10-01 20:28:46.747982] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:51.612 [2024-10-01 20:28:46.748043] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:29:51.612 BaseBdev3 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:51.612 spare_malloc 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:51.612 spare_delay 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:51.612 [2024-10-01 20:28:46.810556] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:29:51.612 [2024-10-01 20:28:46.810629] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:29:51.612 [2024-10-01 20:28:46.810661] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:29:51.612 [2024-10-01 20:28:46.810684] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:29:51.612 [2024-10-01 20:28:46.814017] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:29:51.612 [2024-10-01 20:28:46.814089] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:29:51.612 spare 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:51.612 [2024-10-01 20:28:46.818839] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:29:51.612 [2024-10-01 20:28:46.821823] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:29:51.612 [2024-10-01 20:28:46.821937] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:29:51.612 [2024-10-01 20:28:46.822224] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:29:51.612 [2024-10-01 20:28:46.822257] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:29:51.612 [2024-10-01 20:28:46.822602] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:29:51.612 [2024-10-01 20:28:46.828451] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:29:51.612 [2024-10-01 20:28:46.828515] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:29:51.612 [2024-10-01 20:28:46.828861] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:51.612 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:51.872 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:51.872 "name": "raid_bdev1", 00:29:51.872 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:29:51.872 "strip_size_kb": 64, 00:29:51.872 "state": "online", 00:29:51.872 "raid_level": "raid5f", 00:29:51.872 "superblock": true, 00:29:51.872 "num_base_bdevs": 3, 00:29:51.872 "num_base_bdevs_discovered": 3, 00:29:51.872 "num_base_bdevs_operational": 3, 00:29:51.872 "base_bdevs_list": [ 00:29:51.872 { 00:29:51.872 "name": "BaseBdev1", 00:29:51.872 "uuid": "b742cbe2-343f-59ad-880f-2505530f8cc9", 00:29:51.872 "is_configured": true, 00:29:51.872 "data_offset": 2048, 00:29:51.872 "data_size": 63488 00:29:51.872 }, 00:29:51.872 { 00:29:51.872 "name": "BaseBdev2", 00:29:51.872 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:29:51.872 "is_configured": true, 00:29:51.872 "data_offset": 2048, 00:29:51.872 "data_size": 63488 00:29:51.872 }, 00:29:51.872 { 00:29:51.872 "name": "BaseBdev3", 00:29:51.872 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:29:51.872 "is_configured": true, 00:29:51.872 "data_offset": 2048, 00:29:51.872 "data_size": 63488 00:29:51.872 } 00:29:51.872 ] 00:29:51.872 }' 00:29:51.872 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:51.872 20:28:46 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:52.131 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:29:52.131 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:29:52.131 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:52.131 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:52.131 [2024-10-01 20:28:47.347097] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:29:52.131 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:52.131 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=126976 00:29:52.131 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:29:52.131 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:29:52.391 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:29:52.651 [2024-10-01 20:28:47.714931] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:29:52.651 /dev/nbd0 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:29:52.651 1+0 records in 00:29:52.651 1+0 records out 00:29:52.651 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00034083 s, 12.0 MB/s 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 128 00:29:52.651 20:28:47 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=496 oflag=direct 00:29:53.220 496+0 records in 00:29:53.220 496+0 records out 00:29:53.220 65011712 bytes (65 MB, 62 MiB) copied, 0.498011 s, 131 MB/s 00:29:53.220 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:29:53.220 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:29:53.220 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:29:53.220 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:29:53.220 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:29:53.220 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:29:53.220 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:29:53.479 [2024-10-01 20:28:48.576655] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:53.479 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:29:53.479 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:29:53.479 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:29:53.479 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:29:53.479 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:29:53.479 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:29:53.479 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:29:53.479 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:29:53.479 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:29:53.479 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:53.479 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:53.479 [2024-10-01 20:28:48.594489] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:29:53.479 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:53.480 "name": "raid_bdev1", 00:29:53.480 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:29:53.480 "strip_size_kb": 64, 00:29:53.480 "state": "online", 00:29:53.480 "raid_level": "raid5f", 00:29:53.480 "superblock": true, 00:29:53.480 "num_base_bdevs": 3, 00:29:53.480 "num_base_bdevs_discovered": 2, 00:29:53.480 "num_base_bdevs_operational": 2, 00:29:53.480 "base_bdevs_list": [ 00:29:53.480 { 00:29:53.480 "name": null, 00:29:53.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:53.480 "is_configured": false, 00:29:53.480 "data_offset": 0, 00:29:53.480 "data_size": 63488 00:29:53.480 }, 00:29:53.480 { 00:29:53.480 "name": "BaseBdev2", 00:29:53.480 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:29:53.480 "is_configured": true, 00:29:53.480 "data_offset": 2048, 00:29:53.480 "data_size": 63488 00:29:53.480 }, 00:29:53.480 { 00:29:53.480 "name": "BaseBdev3", 00:29:53.480 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:29:53.480 "is_configured": true, 00:29:53.480 "data_offset": 2048, 00:29:53.480 "data_size": 63488 00:29:53.480 } 00:29:53.480 ] 00:29:53.480 }' 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:53.480 20:28:48 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:54.048 20:28:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:29:54.048 20:28:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:54.048 20:28:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:54.048 [2024-10-01 20:28:49.091281] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:29:54.048 [2024-10-01 20:28:49.108248] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000028f80 00:29:54.048 20:28:49 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:54.048 20:28:49 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:29:54.048 [2024-10-01 20:28:49.116465] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:29:54.986 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:54.986 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:29:54.986 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:29:54.986 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:29:54.986 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:29:54.986 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:54.986 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:54.986 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:54.986 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:54.986 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:54.986 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:29:54.986 "name": "raid_bdev1", 00:29:54.986 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:29:54.986 "strip_size_kb": 64, 00:29:54.986 "state": "online", 00:29:54.986 "raid_level": "raid5f", 00:29:54.986 "superblock": true, 00:29:54.986 "num_base_bdevs": 3, 00:29:54.986 "num_base_bdevs_discovered": 3, 00:29:54.986 "num_base_bdevs_operational": 3, 00:29:54.986 "process": { 00:29:54.986 "type": "rebuild", 00:29:54.986 "target": "spare", 00:29:54.986 "progress": { 00:29:54.986 "blocks": 18432, 00:29:54.986 "percent": 14 00:29:54.986 } 00:29:54.986 }, 00:29:54.986 "base_bdevs_list": [ 00:29:54.986 { 00:29:54.986 "name": "spare", 00:29:54.986 "uuid": "2bb35cc1-2105-5492-bc20-65376db9a40e", 00:29:54.986 "is_configured": true, 00:29:54.986 "data_offset": 2048, 00:29:54.986 "data_size": 63488 00:29:54.986 }, 00:29:54.986 { 00:29:54.986 "name": "BaseBdev2", 00:29:54.986 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:29:54.986 "is_configured": true, 00:29:54.986 "data_offset": 2048, 00:29:54.986 "data_size": 63488 00:29:54.986 }, 00:29:54.986 { 00:29:54.986 "name": "BaseBdev3", 00:29:54.986 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:29:54.986 "is_configured": true, 00:29:54.986 "data_offset": 2048, 00:29:54.986 "data_size": 63488 00:29:54.986 } 00:29:54.986 ] 00:29:54.986 }' 00:29:54.986 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:29:54.986 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:54.986 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:55.247 [2024-10-01 20:28:50.280220] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:29:55.247 [2024-10-01 20:28:50.333281] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:29:55.247 [2024-10-01 20:28:50.333400] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:29:55.247 [2024-10-01 20:28:50.333436] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:29:55.247 [2024-10-01 20:28:50.333451] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:29:55.247 "name": "raid_bdev1", 00:29:55.247 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:29:55.247 "strip_size_kb": 64, 00:29:55.247 "state": "online", 00:29:55.247 "raid_level": "raid5f", 00:29:55.247 "superblock": true, 00:29:55.247 "num_base_bdevs": 3, 00:29:55.247 "num_base_bdevs_discovered": 2, 00:29:55.247 "num_base_bdevs_operational": 2, 00:29:55.247 "base_bdevs_list": [ 00:29:55.247 { 00:29:55.247 "name": null, 00:29:55.247 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:55.247 "is_configured": false, 00:29:55.247 "data_offset": 0, 00:29:55.247 "data_size": 63488 00:29:55.247 }, 00:29:55.247 { 00:29:55.247 "name": "BaseBdev2", 00:29:55.247 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:29:55.247 "is_configured": true, 00:29:55.247 "data_offset": 2048, 00:29:55.247 "data_size": 63488 00:29:55.247 }, 00:29:55.247 { 00:29:55.247 "name": "BaseBdev3", 00:29:55.247 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:29:55.247 "is_configured": true, 00:29:55.247 "data_offset": 2048, 00:29:55.247 "data_size": 63488 00:29:55.247 } 00:29:55.247 ] 00:29:55.247 }' 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:29:55.247 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:55.819 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:29:55.819 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:29:55.819 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:29:55.819 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:29:55.819 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:29:55.819 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:55.819 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:55.819 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:55.819 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:55.819 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:55.819 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:29:55.819 "name": "raid_bdev1", 00:29:55.819 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:29:55.819 "strip_size_kb": 64, 00:29:55.819 "state": "online", 00:29:55.819 "raid_level": "raid5f", 00:29:55.819 "superblock": true, 00:29:55.819 "num_base_bdevs": 3, 00:29:55.819 "num_base_bdevs_discovered": 2, 00:29:55.819 "num_base_bdevs_operational": 2, 00:29:55.819 "base_bdevs_list": [ 00:29:55.819 { 00:29:55.819 "name": null, 00:29:55.819 "uuid": "00000000-0000-0000-0000-000000000000", 00:29:55.819 "is_configured": false, 00:29:55.819 "data_offset": 0, 00:29:55.819 "data_size": 63488 00:29:55.819 }, 00:29:55.819 { 00:29:55.819 "name": "BaseBdev2", 00:29:55.819 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:29:55.819 "is_configured": true, 00:29:55.819 "data_offset": 2048, 00:29:55.819 "data_size": 63488 00:29:55.819 }, 00:29:55.819 { 00:29:55.819 "name": "BaseBdev3", 00:29:55.819 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:29:55.819 "is_configured": true, 00:29:55.819 "data_offset": 2048, 00:29:55.819 "data_size": 63488 00:29:55.819 } 00:29:55.819 ] 00:29:55.819 }' 00:29:55.819 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:29:55.819 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:29:55.819 20:28:50 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:29:55.819 20:28:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:29:55.819 20:28:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:29:55.819 20:28:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:55.819 20:28:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:55.819 [2024-10-01 20:28:51.051796] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:29:56.078 [2024-10-01 20:28:51.071383] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000029050 00:29:56.078 20:28:51 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:56.078 20:28:51 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:29:56.078 [2024-10-01 20:28:51.079921] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:29:57.017 "name": "raid_bdev1", 00:29:57.017 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:29:57.017 "strip_size_kb": 64, 00:29:57.017 "state": "online", 00:29:57.017 "raid_level": "raid5f", 00:29:57.017 "superblock": true, 00:29:57.017 "num_base_bdevs": 3, 00:29:57.017 "num_base_bdevs_discovered": 3, 00:29:57.017 "num_base_bdevs_operational": 3, 00:29:57.017 "process": { 00:29:57.017 "type": "rebuild", 00:29:57.017 "target": "spare", 00:29:57.017 "progress": { 00:29:57.017 "blocks": 18432, 00:29:57.017 "percent": 14 00:29:57.017 } 00:29:57.017 }, 00:29:57.017 "base_bdevs_list": [ 00:29:57.017 { 00:29:57.017 "name": "spare", 00:29:57.017 "uuid": "2bb35cc1-2105-5492-bc20-65376db9a40e", 00:29:57.017 "is_configured": true, 00:29:57.017 "data_offset": 2048, 00:29:57.017 "data_size": 63488 00:29:57.017 }, 00:29:57.017 { 00:29:57.017 "name": "BaseBdev2", 00:29:57.017 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:29:57.017 "is_configured": true, 00:29:57.017 "data_offset": 2048, 00:29:57.017 "data_size": 63488 00:29:57.017 }, 00:29:57.017 { 00:29:57.017 "name": "BaseBdev3", 00:29:57.017 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:29:57.017 "is_configured": true, 00:29:57.017 "data_offset": 2048, 00:29:57.017 "data_size": 63488 00:29:57.017 } 00:29:57.017 ] 00:29:57.017 }' 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:29:57.017 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=671 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:57.017 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:57.277 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:29:57.277 "name": "raid_bdev1", 00:29:57.277 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:29:57.277 "strip_size_kb": 64, 00:29:57.277 "state": "online", 00:29:57.277 "raid_level": "raid5f", 00:29:57.277 "superblock": true, 00:29:57.277 "num_base_bdevs": 3, 00:29:57.277 "num_base_bdevs_discovered": 3, 00:29:57.277 "num_base_bdevs_operational": 3, 00:29:57.277 "process": { 00:29:57.277 "type": "rebuild", 00:29:57.277 "target": "spare", 00:29:57.277 "progress": { 00:29:57.277 "blocks": 22528, 00:29:57.277 "percent": 17 00:29:57.277 } 00:29:57.277 }, 00:29:57.277 "base_bdevs_list": [ 00:29:57.277 { 00:29:57.277 "name": "spare", 00:29:57.277 "uuid": "2bb35cc1-2105-5492-bc20-65376db9a40e", 00:29:57.277 "is_configured": true, 00:29:57.277 "data_offset": 2048, 00:29:57.277 "data_size": 63488 00:29:57.277 }, 00:29:57.277 { 00:29:57.277 "name": "BaseBdev2", 00:29:57.277 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:29:57.277 "is_configured": true, 00:29:57.277 "data_offset": 2048, 00:29:57.277 "data_size": 63488 00:29:57.277 }, 00:29:57.277 { 00:29:57.277 "name": "BaseBdev3", 00:29:57.277 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:29:57.277 "is_configured": true, 00:29:57.277 "data_offset": 2048, 00:29:57.277 "data_size": 63488 00:29:57.277 } 00:29:57.277 ] 00:29:57.277 }' 00:29:57.277 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:29:57.277 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:57.277 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:29:57.277 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:29:57.277 20:28:52 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:29:58.214 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:29:58.214 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:58.214 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:29:58.214 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:29:58.214 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:29:58.214 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:29:58.214 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:58.214 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:58.214 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:58.214 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:58.214 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:58.214 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:29:58.214 "name": "raid_bdev1", 00:29:58.214 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:29:58.214 "strip_size_kb": 64, 00:29:58.214 "state": "online", 00:29:58.214 "raid_level": "raid5f", 00:29:58.214 "superblock": true, 00:29:58.214 "num_base_bdevs": 3, 00:29:58.214 "num_base_bdevs_discovered": 3, 00:29:58.214 "num_base_bdevs_operational": 3, 00:29:58.214 "process": { 00:29:58.214 "type": "rebuild", 00:29:58.214 "target": "spare", 00:29:58.214 "progress": { 00:29:58.214 "blocks": 45056, 00:29:58.214 "percent": 35 00:29:58.214 } 00:29:58.214 }, 00:29:58.214 "base_bdevs_list": [ 00:29:58.214 { 00:29:58.214 "name": "spare", 00:29:58.214 "uuid": "2bb35cc1-2105-5492-bc20-65376db9a40e", 00:29:58.214 "is_configured": true, 00:29:58.214 "data_offset": 2048, 00:29:58.214 "data_size": 63488 00:29:58.214 }, 00:29:58.214 { 00:29:58.214 "name": "BaseBdev2", 00:29:58.214 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:29:58.214 "is_configured": true, 00:29:58.214 "data_offset": 2048, 00:29:58.214 "data_size": 63488 00:29:58.214 }, 00:29:58.214 { 00:29:58.214 "name": "BaseBdev3", 00:29:58.214 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:29:58.214 "is_configured": true, 00:29:58.214 "data_offset": 2048, 00:29:58.214 "data_size": 63488 00:29:58.214 } 00:29:58.214 ] 00:29:58.214 }' 00:29:58.214 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:29:58.472 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:58.472 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:29:58.472 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:29:58.472 20:28:53 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:29:59.404 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:29:59.404 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:29:59.404 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:29:59.404 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:29:59.404 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:29:59.405 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:29:59.405 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:29:59.405 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:29:59.405 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:59.405 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:29:59.405 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:59.405 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:29:59.405 "name": "raid_bdev1", 00:29:59.405 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:29:59.405 "strip_size_kb": 64, 00:29:59.405 "state": "online", 00:29:59.405 "raid_level": "raid5f", 00:29:59.405 "superblock": true, 00:29:59.405 "num_base_bdevs": 3, 00:29:59.405 "num_base_bdevs_discovered": 3, 00:29:59.405 "num_base_bdevs_operational": 3, 00:29:59.405 "process": { 00:29:59.405 "type": "rebuild", 00:29:59.405 "target": "spare", 00:29:59.405 "progress": { 00:29:59.405 "blocks": 69632, 00:29:59.405 "percent": 54 00:29:59.405 } 00:29:59.405 }, 00:29:59.405 "base_bdevs_list": [ 00:29:59.405 { 00:29:59.405 "name": "spare", 00:29:59.405 "uuid": "2bb35cc1-2105-5492-bc20-65376db9a40e", 00:29:59.405 "is_configured": true, 00:29:59.405 "data_offset": 2048, 00:29:59.405 "data_size": 63488 00:29:59.405 }, 00:29:59.405 { 00:29:59.405 "name": "BaseBdev2", 00:29:59.405 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:29:59.405 "is_configured": true, 00:29:59.405 "data_offset": 2048, 00:29:59.405 "data_size": 63488 00:29:59.405 }, 00:29:59.405 { 00:29:59.405 "name": "BaseBdev3", 00:29:59.405 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:29:59.405 "is_configured": true, 00:29:59.405 "data_offset": 2048, 00:29:59.405 "data_size": 63488 00:29:59.405 } 00:29:59.405 ] 00:29:59.405 }' 00:29:59.405 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:29:59.662 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:29:59.662 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:29:59.662 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:29:59.662 20:28:54 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:30:00.594 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:30:00.594 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:00.594 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:00.594 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:00.594 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:00.594 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:00.594 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:00.594 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:00.594 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:00.594 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:00.594 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:00.594 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:00.594 "name": "raid_bdev1", 00:30:00.594 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:00.594 "strip_size_kb": 64, 00:30:00.594 "state": "online", 00:30:00.594 "raid_level": "raid5f", 00:30:00.594 "superblock": true, 00:30:00.594 "num_base_bdevs": 3, 00:30:00.594 "num_base_bdevs_discovered": 3, 00:30:00.594 "num_base_bdevs_operational": 3, 00:30:00.594 "process": { 00:30:00.594 "type": "rebuild", 00:30:00.594 "target": "spare", 00:30:00.594 "progress": { 00:30:00.594 "blocks": 92160, 00:30:00.594 "percent": 72 00:30:00.594 } 00:30:00.594 }, 00:30:00.594 "base_bdevs_list": [ 00:30:00.594 { 00:30:00.594 "name": "spare", 00:30:00.594 "uuid": "2bb35cc1-2105-5492-bc20-65376db9a40e", 00:30:00.594 "is_configured": true, 00:30:00.594 "data_offset": 2048, 00:30:00.594 "data_size": 63488 00:30:00.594 }, 00:30:00.594 { 00:30:00.594 "name": "BaseBdev2", 00:30:00.594 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:00.594 "is_configured": true, 00:30:00.594 "data_offset": 2048, 00:30:00.594 "data_size": 63488 00:30:00.594 }, 00:30:00.594 { 00:30:00.594 "name": "BaseBdev3", 00:30:00.594 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:00.594 "is_configured": true, 00:30:00.594 "data_offset": 2048, 00:30:00.594 "data_size": 63488 00:30:00.594 } 00:30:00.594 ] 00:30:00.594 }' 00:30:00.594 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:00.594 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:00.594 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:00.852 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:00.852 20:28:55 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:30:01.804 20:28:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:30:01.804 20:28:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:01.804 20:28:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:01.804 20:28:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:01.804 20:28:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:01.804 20:28:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:01.804 20:28:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:01.804 20:28:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:01.804 20:28:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:01.804 20:28:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:01.804 20:28:56 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:01.804 20:28:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:01.804 "name": "raid_bdev1", 00:30:01.804 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:01.804 "strip_size_kb": 64, 00:30:01.804 "state": "online", 00:30:01.804 "raid_level": "raid5f", 00:30:01.804 "superblock": true, 00:30:01.804 "num_base_bdevs": 3, 00:30:01.804 "num_base_bdevs_discovered": 3, 00:30:01.804 "num_base_bdevs_operational": 3, 00:30:01.804 "process": { 00:30:01.804 "type": "rebuild", 00:30:01.804 "target": "spare", 00:30:01.804 "progress": { 00:30:01.804 "blocks": 116736, 00:30:01.804 "percent": 91 00:30:01.804 } 00:30:01.804 }, 00:30:01.804 "base_bdevs_list": [ 00:30:01.804 { 00:30:01.804 "name": "spare", 00:30:01.804 "uuid": "2bb35cc1-2105-5492-bc20-65376db9a40e", 00:30:01.804 "is_configured": true, 00:30:01.804 "data_offset": 2048, 00:30:01.804 "data_size": 63488 00:30:01.804 }, 00:30:01.804 { 00:30:01.804 "name": "BaseBdev2", 00:30:01.804 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:01.804 "is_configured": true, 00:30:01.804 "data_offset": 2048, 00:30:01.804 "data_size": 63488 00:30:01.804 }, 00:30:01.804 { 00:30:01.804 "name": "BaseBdev3", 00:30:01.804 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:01.804 "is_configured": true, 00:30:01.804 "data_offset": 2048, 00:30:01.804 "data_size": 63488 00:30:01.804 } 00:30:01.804 ] 00:30:01.804 }' 00:30:01.804 20:28:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:01.804 20:28:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:01.804 20:28:56 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:01.804 20:28:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:01.804 20:28:57 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:30:02.372 [2024-10-01 20:28:57.361029] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:30:02.372 [2024-10-01 20:28:57.361306] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:30:02.372 [2024-10-01 20:28:57.361544] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:02.940 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:30:02.940 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:02.940 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:02.940 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:02.940 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:02.940 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:02.941 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:02.941 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:02.941 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:02.941 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:02.941 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:02.941 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:02.941 "name": "raid_bdev1", 00:30:02.941 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:02.941 "strip_size_kb": 64, 00:30:02.941 "state": "online", 00:30:02.941 "raid_level": "raid5f", 00:30:02.941 "superblock": true, 00:30:02.941 "num_base_bdevs": 3, 00:30:02.941 "num_base_bdevs_discovered": 3, 00:30:02.941 "num_base_bdevs_operational": 3, 00:30:02.941 "base_bdevs_list": [ 00:30:02.941 { 00:30:02.941 "name": "spare", 00:30:02.941 "uuid": "2bb35cc1-2105-5492-bc20-65376db9a40e", 00:30:02.941 "is_configured": true, 00:30:02.941 "data_offset": 2048, 00:30:02.941 "data_size": 63488 00:30:02.941 }, 00:30:02.941 { 00:30:02.941 "name": "BaseBdev2", 00:30:02.941 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:02.941 "is_configured": true, 00:30:02.941 "data_offset": 2048, 00:30:02.941 "data_size": 63488 00:30:02.941 }, 00:30:02.941 { 00:30:02.941 "name": "BaseBdev3", 00:30:02.941 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:02.941 "is_configured": true, 00:30:02.941 "data_offset": 2048, 00:30:02.941 "data_size": 63488 00:30:02.941 } 00:30:02.941 ] 00:30:02.941 }' 00:30:02.941 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:02.941 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:30:02.941 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:03.200 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:30:03.200 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:30:03.200 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:30:03.200 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:03.200 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:30:03.200 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:30:03.200 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:03.200 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:03.200 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:03.200 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:03.200 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.200 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:03.200 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:03.200 "name": "raid_bdev1", 00:30:03.200 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:03.200 "strip_size_kb": 64, 00:30:03.200 "state": "online", 00:30:03.200 "raid_level": "raid5f", 00:30:03.200 "superblock": true, 00:30:03.200 "num_base_bdevs": 3, 00:30:03.200 "num_base_bdevs_discovered": 3, 00:30:03.200 "num_base_bdevs_operational": 3, 00:30:03.200 "base_bdevs_list": [ 00:30:03.200 { 00:30:03.200 "name": "spare", 00:30:03.200 "uuid": "2bb35cc1-2105-5492-bc20-65376db9a40e", 00:30:03.200 "is_configured": true, 00:30:03.200 "data_offset": 2048, 00:30:03.200 "data_size": 63488 00:30:03.200 }, 00:30:03.200 { 00:30:03.200 "name": "BaseBdev2", 00:30:03.200 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:03.200 "is_configured": true, 00:30:03.200 "data_offset": 2048, 00:30:03.200 "data_size": 63488 00:30:03.200 }, 00:30:03.200 { 00:30:03.200 "name": "BaseBdev3", 00:30:03.200 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:03.200 "is_configured": true, 00:30:03.200 "data_offset": 2048, 00:30:03.200 "data_size": 63488 00:30:03.201 } 00:30:03.201 ] 00:30:03.201 }' 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:03.201 "name": "raid_bdev1", 00:30:03.201 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:03.201 "strip_size_kb": 64, 00:30:03.201 "state": "online", 00:30:03.201 "raid_level": "raid5f", 00:30:03.201 "superblock": true, 00:30:03.201 "num_base_bdevs": 3, 00:30:03.201 "num_base_bdevs_discovered": 3, 00:30:03.201 "num_base_bdevs_operational": 3, 00:30:03.201 "base_bdevs_list": [ 00:30:03.201 { 00:30:03.201 "name": "spare", 00:30:03.201 "uuid": "2bb35cc1-2105-5492-bc20-65376db9a40e", 00:30:03.201 "is_configured": true, 00:30:03.201 "data_offset": 2048, 00:30:03.201 "data_size": 63488 00:30:03.201 }, 00:30:03.201 { 00:30:03.201 "name": "BaseBdev2", 00:30:03.201 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:03.201 "is_configured": true, 00:30:03.201 "data_offset": 2048, 00:30:03.201 "data_size": 63488 00:30:03.201 }, 00:30:03.201 { 00:30:03.201 "name": "BaseBdev3", 00:30:03.201 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:03.201 "is_configured": true, 00:30:03.201 "data_offset": 2048, 00:30:03.201 "data_size": 63488 00:30:03.201 } 00:30:03.201 ] 00:30:03.201 }' 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:03.201 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.767 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:30:03.767 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:03.767 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.768 [2024-10-01 20:28:58.869303] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:30:03.768 [2024-10-01 20:28:58.869379] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:03.768 [2024-10-01 20:28:58.869571] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:03.768 [2024-10-01 20:28:58.869688] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:03.768 [2024-10-01 20:28:58.869737] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:30:03.768 20:28:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:30:04.026 /dev/nbd0 00:30:04.026 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:30:04.026 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:30:04.026 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:30:04.026 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:30:04.026 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:30:04.026 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:30:04.026 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:30:04.284 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:30:04.284 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:30:04.284 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:30:04.284 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:30:04.284 1+0 records in 00:30:04.284 1+0 records out 00:30:04.284 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000399172 s, 10.3 MB/s 00:30:04.284 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:04.284 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:30:04.284 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:04.284 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:30:04.284 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:30:04.284 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:30:04.284 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:30:04.284 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:30:04.543 /dev/nbd1 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:30:04.543 1+0 records in 00:30:04.543 1+0 records out 00:30:04.543 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00047109 s, 8.7 MB/s 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:30:04.543 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:30:04.802 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:30:04.802 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:30:04.802 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:30:04.802 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:30:04.802 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:30:04.802 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:30:04.802 20:28:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:30:05.061 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:30:05.061 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:30:05.061 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:30:05.061 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:30:05.061 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:30:05.061 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:30:05.061 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:30:05.061 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:30:05.061 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:30:05.061 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:30:05.319 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:30:05.319 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:30:05.319 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:30:05.319 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:30:05.319 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:30:05.319 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:30:05.319 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:30:05.319 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:30:05.319 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:30:05.319 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:30:05.319 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:05.319 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:05.319 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:05.320 [2024-10-01 20:29:00.435385] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:30:05.320 [2024-10-01 20:29:00.435463] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:05.320 [2024-10-01 20:29:00.435514] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:30:05.320 [2024-10-01 20:29:00.435547] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:05.320 [2024-10-01 20:29:00.439249] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:05.320 [2024-10-01 20:29:00.439340] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:30:05.320 [2024-10-01 20:29:00.439478] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:30:05.320 [2024-10-01 20:29:00.439587] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:30:05.320 [2024-10-01 20:29:00.439898] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:05.320 spare 00:30:05.320 [2024-10-01 20:29:00.440095] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:05.320 [2024-10-01 20:29:00.540252] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:30:05.320 [2024-10-01 20:29:00.540316] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:30:05.320 [2024-10-01 20:29:00.540737] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000047700 00:30:05.320 [2024-10-01 20:29:00.545933] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:30:05.320 [2024-10-01 20:29:00.545969] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:30:05.320 [2024-10-01 20:29:00.546240] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:05.320 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:05.578 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:05.578 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:05.578 "name": "raid_bdev1", 00:30:05.578 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:05.578 "strip_size_kb": 64, 00:30:05.578 "state": "online", 00:30:05.578 "raid_level": "raid5f", 00:30:05.578 "superblock": true, 00:30:05.578 "num_base_bdevs": 3, 00:30:05.578 "num_base_bdevs_discovered": 3, 00:30:05.578 "num_base_bdevs_operational": 3, 00:30:05.578 "base_bdevs_list": [ 00:30:05.578 { 00:30:05.578 "name": "spare", 00:30:05.578 "uuid": "2bb35cc1-2105-5492-bc20-65376db9a40e", 00:30:05.578 "is_configured": true, 00:30:05.578 "data_offset": 2048, 00:30:05.578 "data_size": 63488 00:30:05.578 }, 00:30:05.578 { 00:30:05.578 "name": "BaseBdev2", 00:30:05.578 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:05.578 "is_configured": true, 00:30:05.578 "data_offset": 2048, 00:30:05.578 "data_size": 63488 00:30:05.578 }, 00:30:05.578 { 00:30:05.578 "name": "BaseBdev3", 00:30:05.578 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:05.579 "is_configured": true, 00:30:05.579 "data_offset": 2048, 00:30:05.579 "data_size": 63488 00:30:05.579 } 00:30:05.579 ] 00:30:05.579 }' 00:30:05.579 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:05.579 20:29:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:05.836 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:30:05.836 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:05.836 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:30:05.836 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:30:05.837 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:05.837 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:05.837 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:05.837 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:05.837 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:06.095 "name": "raid_bdev1", 00:30:06.095 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:06.095 "strip_size_kb": 64, 00:30:06.095 "state": "online", 00:30:06.095 "raid_level": "raid5f", 00:30:06.095 "superblock": true, 00:30:06.095 "num_base_bdevs": 3, 00:30:06.095 "num_base_bdevs_discovered": 3, 00:30:06.095 "num_base_bdevs_operational": 3, 00:30:06.095 "base_bdevs_list": [ 00:30:06.095 { 00:30:06.095 "name": "spare", 00:30:06.095 "uuid": "2bb35cc1-2105-5492-bc20-65376db9a40e", 00:30:06.095 "is_configured": true, 00:30:06.095 "data_offset": 2048, 00:30:06.095 "data_size": 63488 00:30:06.095 }, 00:30:06.095 { 00:30:06.095 "name": "BaseBdev2", 00:30:06.095 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:06.095 "is_configured": true, 00:30:06.095 "data_offset": 2048, 00:30:06.095 "data_size": 63488 00:30:06.095 }, 00:30:06.095 { 00:30:06.095 "name": "BaseBdev3", 00:30:06.095 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:06.095 "is_configured": true, 00:30:06.095 "data_offset": 2048, 00:30:06.095 "data_size": 63488 00:30:06.095 } 00:30:06.095 ] 00:30:06.095 }' 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:06.095 [2024-10-01 20:29:01.284138] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:06.095 "name": "raid_bdev1", 00:30:06.095 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:06.095 "strip_size_kb": 64, 00:30:06.095 "state": "online", 00:30:06.095 "raid_level": "raid5f", 00:30:06.095 "superblock": true, 00:30:06.095 "num_base_bdevs": 3, 00:30:06.095 "num_base_bdevs_discovered": 2, 00:30:06.095 "num_base_bdevs_operational": 2, 00:30:06.095 "base_bdevs_list": [ 00:30:06.095 { 00:30:06.095 "name": null, 00:30:06.095 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:06.095 "is_configured": false, 00:30:06.095 "data_offset": 0, 00:30:06.095 "data_size": 63488 00:30:06.095 }, 00:30:06.095 { 00:30:06.095 "name": "BaseBdev2", 00:30:06.095 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:06.095 "is_configured": true, 00:30:06.095 "data_offset": 2048, 00:30:06.095 "data_size": 63488 00:30:06.095 }, 00:30:06.095 { 00:30:06.095 "name": "BaseBdev3", 00:30:06.095 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:06.095 "is_configured": true, 00:30:06.095 "data_offset": 2048, 00:30:06.095 "data_size": 63488 00:30:06.095 } 00:30:06.095 ] 00:30:06.095 }' 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:06.095 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:06.663 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:30:06.663 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:06.663 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:06.663 [2024-10-01 20:29:01.808414] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:30:06.663 [2024-10-01 20:29:01.808679] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:30:06.663 [2024-10-01 20:29:01.808729] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:30:06.663 [2024-10-01 20:29:01.808785] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:30:06.663 [2024-10-01 20:29:01.822937] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000477d0 00:30:06.663 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:06.663 20:29:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:30:06.663 [2024-10-01 20:29:01.830467] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:30:07.599 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:07.599 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:07.599 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:07.599 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:07.599 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:07.599 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:07.599 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:07.599 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:07.599 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:07.599 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:07.858 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:07.858 "name": "raid_bdev1", 00:30:07.858 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:07.858 "strip_size_kb": 64, 00:30:07.858 "state": "online", 00:30:07.858 "raid_level": "raid5f", 00:30:07.858 "superblock": true, 00:30:07.858 "num_base_bdevs": 3, 00:30:07.858 "num_base_bdevs_discovered": 3, 00:30:07.858 "num_base_bdevs_operational": 3, 00:30:07.858 "process": { 00:30:07.858 "type": "rebuild", 00:30:07.858 "target": "spare", 00:30:07.858 "progress": { 00:30:07.858 "blocks": 18432, 00:30:07.858 "percent": 14 00:30:07.858 } 00:30:07.858 }, 00:30:07.858 "base_bdevs_list": [ 00:30:07.858 { 00:30:07.858 "name": "spare", 00:30:07.858 "uuid": "2bb35cc1-2105-5492-bc20-65376db9a40e", 00:30:07.858 "is_configured": true, 00:30:07.858 "data_offset": 2048, 00:30:07.858 "data_size": 63488 00:30:07.858 }, 00:30:07.858 { 00:30:07.858 "name": "BaseBdev2", 00:30:07.858 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:07.858 "is_configured": true, 00:30:07.858 "data_offset": 2048, 00:30:07.858 "data_size": 63488 00:30:07.858 }, 00:30:07.858 { 00:30:07.858 "name": "BaseBdev3", 00:30:07.858 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:07.858 "is_configured": true, 00:30:07.858 "data_offset": 2048, 00:30:07.858 "data_size": 63488 00:30:07.858 } 00:30:07.858 ] 00:30:07.858 }' 00:30:07.858 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:07.858 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:07.858 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:07.858 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:07.858 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:30:07.858 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:07.858 20:29:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:07.858 [2024-10-01 20:29:02.985167] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:30:07.858 [2024-10-01 20:29:03.045474] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:30:07.858 [2024-10-01 20:29:03.045586] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:07.858 [2024-10-01 20:29:03.045617] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:30:07.858 [2024-10-01 20:29:03.045636] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:30:07.858 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:07.858 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:30:07.858 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:07.858 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:07.858 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:07.858 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:07.858 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:30:07.858 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:07.858 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:07.858 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:07.858 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:07.858 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:07.858 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:07.858 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:07.858 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:07.858 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:08.116 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:08.117 "name": "raid_bdev1", 00:30:08.117 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:08.117 "strip_size_kb": 64, 00:30:08.117 "state": "online", 00:30:08.117 "raid_level": "raid5f", 00:30:08.117 "superblock": true, 00:30:08.117 "num_base_bdevs": 3, 00:30:08.117 "num_base_bdevs_discovered": 2, 00:30:08.117 "num_base_bdevs_operational": 2, 00:30:08.117 "base_bdevs_list": [ 00:30:08.117 { 00:30:08.117 "name": null, 00:30:08.117 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:08.117 "is_configured": false, 00:30:08.117 "data_offset": 0, 00:30:08.117 "data_size": 63488 00:30:08.117 }, 00:30:08.117 { 00:30:08.117 "name": "BaseBdev2", 00:30:08.117 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:08.117 "is_configured": true, 00:30:08.117 "data_offset": 2048, 00:30:08.117 "data_size": 63488 00:30:08.117 }, 00:30:08.117 { 00:30:08.117 "name": "BaseBdev3", 00:30:08.117 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:08.117 "is_configured": true, 00:30:08.117 "data_offset": 2048, 00:30:08.117 "data_size": 63488 00:30:08.117 } 00:30:08.117 ] 00:30:08.117 }' 00:30:08.117 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:08.117 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:08.375 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:30:08.375 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:08.375 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:08.375 [2024-10-01 20:29:03.590512] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:30:08.375 [2024-10-01 20:29:03.590611] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:08.375 [2024-10-01 20:29:03.590651] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:30:08.375 [2024-10-01 20:29:03.590678] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:08.375 [2024-10-01 20:29:03.591352] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:08.375 [2024-10-01 20:29:03.591402] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:30:08.375 [2024-10-01 20:29:03.591544] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:30:08.375 [2024-10-01 20:29:03.591576] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:30:08.375 [2024-10-01 20:29:03.591593] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:30:08.375 [2024-10-01 20:29:03.591643] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:30:08.375 [2024-10-01 20:29:03.605808] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000478a0 00:30:08.375 spare 00:30:08.375 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:08.375 20:29:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:30:08.375 [2024-10-01 20:29:03.613505] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:30:09.749 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:30:09.749 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:09.749 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:30:09.749 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:30:09.749 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:09.749 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:09.749 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:09.749 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:09.749 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:09.749 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:09.749 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:09.749 "name": "raid_bdev1", 00:30:09.749 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:09.749 "strip_size_kb": 64, 00:30:09.749 "state": "online", 00:30:09.749 "raid_level": "raid5f", 00:30:09.749 "superblock": true, 00:30:09.749 "num_base_bdevs": 3, 00:30:09.749 "num_base_bdevs_discovered": 3, 00:30:09.749 "num_base_bdevs_operational": 3, 00:30:09.749 "process": { 00:30:09.749 "type": "rebuild", 00:30:09.749 "target": "spare", 00:30:09.749 "progress": { 00:30:09.749 "blocks": 18432, 00:30:09.749 "percent": 14 00:30:09.749 } 00:30:09.749 }, 00:30:09.749 "base_bdevs_list": [ 00:30:09.749 { 00:30:09.749 "name": "spare", 00:30:09.749 "uuid": "2bb35cc1-2105-5492-bc20-65376db9a40e", 00:30:09.749 "is_configured": true, 00:30:09.749 "data_offset": 2048, 00:30:09.749 "data_size": 63488 00:30:09.749 }, 00:30:09.749 { 00:30:09.749 "name": "BaseBdev2", 00:30:09.749 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:09.749 "is_configured": true, 00:30:09.749 "data_offset": 2048, 00:30:09.749 "data_size": 63488 00:30:09.749 }, 00:30:09.749 { 00:30:09.749 "name": "BaseBdev3", 00:30:09.749 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:09.749 "is_configured": true, 00:30:09.749 "data_offset": 2048, 00:30:09.749 "data_size": 63488 00:30:09.749 } 00:30:09.749 ] 00:30:09.749 }' 00:30:09.749 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:09.749 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:30:09.749 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:09.749 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:30:09.749 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:09.750 [2024-10-01 20:29:04.775777] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:30:09.750 [2024-10-01 20:29:04.828035] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:30:09.750 [2024-10-01 20:29:04.828137] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:09.750 [2024-10-01 20:29:04.828177] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:30:09.750 [2024-10-01 20:29:04.828192] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:09.750 "name": "raid_bdev1", 00:30:09.750 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:09.750 "strip_size_kb": 64, 00:30:09.750 "state": "online", 00:30:09.750 "raid_level": "raid5f", 00:30:09.750 "superblock": true, 00:30:09.750 "num_base_bdevs": 3, 00:30:09.750 "num_base_bdevs_discovered": 2, 00:30:09.750 "num_base_bdevs_operational": 2, 00:30:09.750 "base_bdevs_list": [ 00:30:09.750 { 00:30:09.750 "name": null, 00:30:09.750 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:09.750 "is_configured": false, 00:30:09.750 "data_offset": 0, 00:30:09.750 "data_size": 63488 00:30:09.750 }, 00:30:09.750 { 00:30:09.750 "name": "BaseBdev2", 00:30:09.750 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:09.750 "is_configured": true, 00:30:09.750 "data_offset": 2048, 00:30:09.750 "data_size": 63488 00:30:09.750 }, 00:30:09.750 { 00:30:09.750 "name": "BaseBdev3", 00:30:09.750 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:09.750 "is_configured": true, 00:30:09.750 "data_offset": 2048, 00:30:09.750 "data_size": 63488 00:30:09.750 } 00:30:09.750 ] 00:30:09.750 }' 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:09.750 20:29:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:10.316 "name": "raid_bdev1", 00:30:10.316 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:10.316 "strip_size_kb": 64, 00:30:10.316 "state": "online", 00:30:10.316 "raid_level": "raid5f", 00:30:10.316 "superblock": true, 00:30:10.316 "num_base_bdevs": 3, 00:30:10.316 "num_base_bdevs_discovered": 2, 00:30:10.316 "num_base_bdevs_operational": 2, 00:30:10.316 "base_bdevs_list": [ 00:30:10.316 { 00:30:10.316 "name": null, 00:30:10.316 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:10.316 "is_configured": false, 00:30:10.316 "data_offset": 0, 00:30:10.316 "data_size": 63488 00:30:10.316 }, 00:30:10.316 { 00:30:10.316 "name": "BaseBdev2", 00:30:10.316 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:10.316 "is_configured": true, 00:30:10.316 "data_offset": 2048, 00:30:10.316 "data_size": 63488 00:30:10.316 }, 00:30:10.316 { 00:30:10.316 "name": "BaseBdev3", 00:30:10.316 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:10.316 "is_configured": true, 00:30:10.316 "data_offset": 2048, 00:30:10.316 "data_size": 63488 00:30:10.316 } 00:30:10.316 ] 00:30:10.316 }' 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:10.316 [2024-10-01 20:29:05.551423] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:30:10.316 [2024-10-01 20:29:05.551549] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:10.316 [2024-10-01 20:29:05.551595] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:30:10.316 [2024-10-01 20:29:05.551614] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:10.316 [2024-10-01 20:29:05.552263] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:10.316 [2024-10-01 20:29:05.552311] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:30:10.316 [2024-10-01 20:29:05.552432] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:30:10.316 [2024-10-01 20:29:05.552457] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:30:10.316 [2024-10-01 20:29:05.552477] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:30:10.316 [2024-10-01 20:29:05.552498] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:30:10.316 BaseBdev1 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:10.316 20:29:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:11.691 "name": "raid_bdev1", 00:30:11.691 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:11.691 "strip_size_kb": 64, 00:30:11.691 "state": "online", 00:30:11.691 "raid_level": "raid5f", 00:30:11.691 "superblock": true, 00:30:11.691 "num_base_bdevs": 3, 00:30:11.691 "num_base_bdevs_discovered": 2, 00:30:11.691 "num_base_bdevs_operational": 2, 00:30:11.691 "base_bdevs_list": [ 00:30:11.691 { 00:30:11.691 "name": null, 00:30:11.691 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:11.691 "is_configured": false, 00:30:11.691 "data_offset": 0, 00:30:11.691 "data_size": 63488 00:30:11.691 }, 00:30:11.691 { 00:30:11.691 "name": "BaseBdev2", 00:30:11.691 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:11.691 "is_configured": true, 00:30:11.691 "data_offset": 2048, 00:30:11.691 "data_size": 63488 00:30:11.691 }, 00:30:11.691 { 00:30:11.691 "name": "BaseBdev3", 00:30:11.691 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:11.691 "is_configured": true, 00:30:11.691 "data_offset": 2048, 00:30:11.691 "data_size": 63488 00:30:11.691 } 00:30:11.691 ] 00:30:11.691 }' 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:11.691 20:29:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:11.950 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:30:11.950 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:11.950 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:30:11.950 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:30:11.950 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:11.950 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:11.950 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:11.950 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:11.950 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:11.950 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:11.950 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:11.950 "name": "raid_bdev1", 00:30:11.950 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:11.950 "strip_size_kb": 64, 00:30:11.950 "state": "online", 00:30:11.950 "raid_level": "raid5f", 00:30:11.950 "superblock": true, 00:30:11.950 "num_base_bdevs": 3, 00:30:11.950 "num_base_bdevs_discovered": 2, 00:30:11.950 "num_base_bdevs_operational": 2, 00:30:11.950 "base_bdevs_list": [ 00:30:11.950 { 00:30:11.950 "name": null, 00:30:11.950 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:11.950 "is_configured": false, 00:30:11.950 "data_offset": 0, 00:30:11.950 "data_size": 63488 00:30:11.950 }, 00:30:11.950 { 00:30:11.950 "name": "BaseBdev2", 00:30:11.950 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:11.950 "is_configured": true, 00:30:11.950 "data_offset": 2048, 00:30:11.950 "data_size": 63488 00:30:11.950 }, 00:30:11.950 { 00:30:11.950 "name": "BaseBdev3", 00:30:11.950 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:11.950 "is_configured": true, 00:30:11.950 "data_offset": 2048, 00:30:11.950 "data_size": 63488 00:30:11.950 } 00:30:11.950 ] 00:30:11.950 }' 00:30:11.950 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:11.950 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:30:11.950 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:12.209 [2024-10-01 20:29:07.244390] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:30:12.209 [2024-10-01 20:29:07.244699] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:30:12.209 [2024-10-01 20:29:07.244729] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:30:12.209 request: 00:30:12.209 { 00:30:12.209 "base_bdev": "BaseBdev1", 00:30:12.209 "raid_bdev": "raid_bdev1", 00:30:12.209 "method": "bdev_raid_add_base_bdev", 00:30:12.209 "req_id": 1 00:30:12.209 } 00:30:12.209 Got JSON-RPC error response 00:30:12.209 response: 00:30:12.209 { 00:30:12.209 "code": -22, 00:30:12.209 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:30:12.209 } 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:30:12.209 20:29:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:30:13.144 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:30:13.144 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:13.144 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:13.144 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:13.144 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:13.144 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:30:13.144 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:13.144 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:13.144 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:13.144 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:13.145 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:13.145 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:13.145 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:13.145 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:13.145 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:13.145 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:13.145 "name": "raid_bdev1", 00:30:13.145 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:13.145 "strip_size_kb": 64, 00:30:13.145 "state": "online", 00:30:13.145 "raid_level": "raid5f", 00:30:13.145 "superblock": true, 00:30:13.145 "num_base_bdevs": 3, 00:30:13.145 "num_base_bdevs_discovered": 2, 00:30:13.145 "num_base_bdevs_operational": 2, 00:30:13.145 "base_bdevs_list": [ 00:30:13.145 { 00:30:13.145 "name": null, 00:30:13.145 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:13.145 "is_configured": false, 00:30:13.145 "data_offset": 0, 00:30:13.145 "data_size": 63488 00:30:13.145 }, 00:30:13.145 { 00:30:13.145 "name": "BaseBdev2", 00:30:13.145 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:13.145 "is_configured": true, 00:30:13.145 "data_offset": 2048, 00:30:13.145 "data_size": 63488 00:30:13.145 }, 00:30:13.145 { 00:30:13.145 "name": "BaseBdev3", 00:30:13.145 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:13.145 "is_configured": true, 00:30:13.145 "data_offset": 2048, 00:30:13.145 "data_size": 63488 00:30:13.145 } 00:30:13.145 ] 00:30:13.145 }' 00:30:13.145 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:13.145 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:30:13.712 "name": "raid_bdev1", 00:30:13.712 "uuid": "fab60925-e47b-41f0-9191-50a4f096e78a", 00:30:13.712 "strip_size_kb": 64, 00:30:13.712 "state": "online", 00:30:13.712 "raid_level": "raid5f", 00:30:13.712 "superblock": true, 00:30:13.712 "num_base_bdevs": 3, 00:30:13.712 "num_base_bdevs_discovered": 2, 00:30:13.712 "num_base_bdevs_operational": 2, 00:30:13.712 "base_bdevs_list": [ 00:30:13.712 { 00:30:13.712 "name": null, 00:30:13.712 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:13.712 "is_configured": false, 00:30:13.712 "data_offset": 0, 00:30:13.712 "data_size": 63488 00:30:13.712 }, 00:30:13.712 { 00:30:13.712 "name": "BaseBdev2", 00:30:13.712 "uuid": "c7593723-c155-5a54-9dde-5c58835c113c", 00:30:13.712 "is_configured": true, 00:30:13.712 "data_offset": 2048, 00:30:13.712 "data_size": 63488 00:30:13.712 }, 00:30:13.712 { 00:30:13.712 "name": "BaseBdev3", 00:30:13.712 "uuid": "88d45a9f-d54d-5d1e-9255-dcc80b08f7b4", 00:30:13.712 "is_configured": true, 00:30:13.712 "data_offset": 2048, 00:30:13.712 "data_size": 63488 00:30:13.712 } 00:30:13.712 ] 00:30:13.712 }' 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 83296 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@950 -- # '[' -z 83296 ']' 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # kill -0 83296 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@955 -- # uname 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:30:13.712 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 83296 00:30:13.971 killing process with pid 83296 00:30:13.971 Received shutdown signal, test time was about 60.000000 seconds 00:30:13.971 00:30:13.971 Latency(us) 00:30:13.971 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:30:13.971 =================================================================================================================== 00:30:13.971 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:30:13.971 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:30:13.971 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:30:13.971 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 83296' 00:30:13.971 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@969 -- # kill 83296 00:30:13.971 [2024-10-01 20:29:08.981565] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:30:13.971 20:29:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@974 -- # wait 83296 00:30:13.971 [2024-10-01 20:29:08.981831] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:13.971 [2024-10-01 20:29:08.981927] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:13.971 [2024-10-01 20:29:08.981953] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:30:14.230 [2024-10-01 20:29:09.344510] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:30:16.197 20:29:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:30:16.197 00:30:16.197 real 0m25.997s 00:30:16.197 user 0m34.039s 00:30:16.197 sys 0m2.883s 00:30:16.197 20:29:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:30:16.197 20:29:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:16.197 ************************************ 00:30:16.197 END TEST raid5f_rebuild_test_sb 00:30:16.197 ************************************ 00:30:16.197 20:29:11 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:30:16.197 20:29:11 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 4 false 00:30:16.197 20:29:11 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:30:16.197 20:29:11 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:30:16.197 20:29:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:30:16.197 ************************************ 00:30:16.197 START TEST raid5f_state_function_test 00:30:16.197 ************************************ 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid5f 4 false 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=84072 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 84072' 00:30:16.198 Process raid pid: 84072 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 84072 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 84072 ']' 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:16.198 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:30:16.198 20:29:11 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:16.457 [2024-10-01 20:29:11.489212] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:30:16.457 [2024-10-01 20:29:11.489405] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:30:16.457 [2024-10-01 20:29:11.665913] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:30:16.716 [2024-10-01 20:29:11.927157] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:30:16.975 [2024-10-01 20:29:12.170304] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:16.975 [2024-10-01 20:29:12.170794] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:17.541 20:29:12 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:30:17.541 20:29:12 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:30:17.541 20:29:12 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:30:17.541 20:29:12 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:17.541 20:29:12 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.541 [2024-10-01 20:29:12.673251] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:30:17.541 [2024-10-01 20:29:12.673416] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:30:17.541 [2024-10-01 20:29:12.673492] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:30:17.541 [2024-10-01 20:29:12.673555] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:30:17.541 [2024-10-01 20:29:12.673603] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:30:17.541 [2024-10-01 20:29:12.673665] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:30:17.541 [2024-10-01 20:29:12.673731] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:30:17.541 [2024-10-01 20:29:12.673797] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:30:17.541 20:29:12 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:17.541 20:29:12 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:17.541 20:29:12 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:17.541 20:29:12 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:17.542 20:29:12 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:17.542 20:29:12 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:17.542 20:29:12 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:17.542 20:29:12 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:17.542 20:29:12 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:17.542 20:29:12 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:17.542 20:29:12 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:17.542 20:29:12 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:17.542 20:29:12 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:17.542 20:29:12 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:17.542 20:29:12 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:17.542 20:29:12 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:17.542 20:29:12 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:17.542 "name": "Existed_Raid", 00:30:17.542 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:17.542 "strip_size_kb": 64, 00:30:17.542 "state": "configuring", 00:30:17.542 "raid_level": "raid5f", 00:30:17.542 "superblock": false, 00:30:17.542 "num_base_bdevs": 4, 00:30:17.542 "num_base_bdevs_discovered": 0, 00:30:17.542 "num_base_bdevs_operational": 4, 00:30:17.542 "base_bdevs_list": [ 00:30:17.542 { 00:30:17.542 "name": "BaseBdev1", 00:30:17.542 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:17.542 "is_configured": false, 00:30:17.542 "data_offset": 0, 00:30:17.542 "data_size": 0 00:30:17.542 }, 00:30:17.542 { 00:30:17.542 "name": "BaseBdev2", 00:30:17.542 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:17.542 "is_configured": false, 00:30:17.542 "data_offset": 0, 00:30:17.542 "data_size": 0 00:30:17.542 }, 00:30:17.542 { 00:30:17.542 "name": "BaseBdev3", 00:30:17.542 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:17.542 "is_configured": false, 00:30:17.542 "data_offset": 0, 00:30:17.542 "data_size": 0 00:30:17.542 }, 00:30:17.542 { 00:30:17.542 "name": "BaseBdev4", 00:30:17.542 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:17.542 "is_configured": false, 00:30:17.542 "data_offset": 0, 00:30:17.542 "data_size": 0 00:30:17.542 } 00:30:17.542 ] 00:30:17.542 }' 00:30:17.542 20:29:12 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:17.542 20:29:12 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:18.109 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:30:18.109 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:18.109 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:18.109 [2024-10-01 20:29:13.189297] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:30:18.110 [2024-10-01 20:29:13.189371] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:18.110 [2024-10-01 20:29:13.197286] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:30:18.110 [2024-10-01 20:29:13.197509] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:30:18.110 [2024-10-01 20:29:13.197540] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:30:18.110 [2024-10-01 20:29:13.197565] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:30:18.110 [2024-10-01 20:29:13.197579] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:30:18.110 [2024-10-01 20:29:13.197599] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:30:18.110 [2024-10-01 20:29:13.197612] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:30:18.110 [2024-10-01 20:29:13.197631] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:18.110 [2024-10-01 20:29:13.250772] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:30:18.110 BaseBdev1 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:18.110 [ 00:30:18.110 { 00:30:18.110 "name": "BaseBdev1", 00:30:18.110 "aliases": [ 00:30:18.110 "cb3a2648-6368-4b77-8e51-355764581eb3" 00:30:18.110 ], 00:30:18.110 "product_name": "Malloc disk", 00:30:18.110 "block_size": 512, 00:30:18.110 "num_blocks": 65536, 00:30:18.110 "uuid": "cb3a2648-6368-4b77-8e51-355764581eb3", 00:30:18.110 "assigned_rate_limits": { 00:30:18.110 "rw_ios_per_sec": 0, 00:30:18.110 "rw_mbytes_per_sec": 0, 00:30:18.110 "r_mbytes_per_sec": 0, 00:30:18.110 "w_mbytes_per_sec": 0 00:30:18.110 }, 00:30:18.110 "claimed": true, 00:30:18.110 "claim_type": "exclusive_write", 00:30:18.110 "zoned": false, 00:30:18.110 "supported_io_types": { 00:30:18.110 "read": true, 00:30:18.110 "write": true, 00:30:18.110 "unmap": true, 00:30:18.110 "flush": true, 00:30:18.110 "reset": true, 00:30:18.110 "nvme_admin": false, 00:30:18.110 "nvme_io": false, 00:30:18.110 "nvme_io_md": false, 00:30:18.110 "write_zeroes": true, 00:30:18.110 "zcopy": true, 00:30:18.110 "get_zone_info": false, 00:30:18.110 "zone_management": false, 00:30:18.110 "zone_append": false, 00:30:18.110 "compare": false, 00:30:18.110 "compare_and_write": false, 00:30:18.110 "abort": true, 00:30:18.110 "seek_hole": false, 00:30:18.110 "seek_data": false, 00:30:18.110 "copy": true, 00:30:18.110 "nvme_iov_md": false 00:30:18.110 }, 00:30:18.110 "memory_domains": [ 00:30:18.110 { 00:30:18.110 "dma_device_id": "system", 00:30:18.110 "dma_device_type": 1 00:30:18.110 }, 00:30:18.110 { 00:30:18.110 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:18.110 "dma_device_type": 2 00:30:18.110 } 00:30:18.110 ], 00:30:18.110 "driver_specific": {} 00:30:18.110 } 00:30:18.110 ] 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:18.110 "name": "Existed_Raid", 00:30:18.110 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:18.110 "strip_size_kb": 64, 00:30:18.110 "state": "configuring", 00:30:18.110 "raid_level": "raid5f", 00:30:18.110 "superblock": false, 00:30:18.110 "num_base_bdevs": 4, 00:30:18.110 "num_base_bdevs_discovered": 1, 00:30:18.110 "num_base_bdevs_operational": 4, 00:30:18.110 "base_bdevs_list": [ 00:30:18.110 { 00:30:18.110 "name": "BaseBdev1", 00:30:18.110 "uuid": "cb3a2648-6368-4b77-8e51-355764581eb3", 00:30:18.110 "is_configured": true, 00:30:18.110 "data_offset": 0, 00:30:18.110 "data_size": 65536 00:30:18.110 }, 00:30:18.110 { 00:30:18.110 "name": "BaseBdev2", 00:30:18.110 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:18.110 "is_configured": false, 00:30:18.110 "data_offset": 0, 00:30:18.110 "data_size": 0 00:30:18.110 }, 00:30:18.110 { 00:30:18.110 "name": "BaseBdev3", 00:30:18.110 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:18.110 "is_configured": false, 00:30:18.110 "data_offset": 0, 00:30:18.110 "data_size": 0 00:30:18.110 }, 00:30:18.110 { 00:30:18.110 "name": "BaseBdev4", 00:30:18.110 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:18.110 "is_configured": false, 00:30:18.110 "data_offset": 0, 00:30:18.110 "data_size": 0 00:30:18.110 } 00:30:18.110 ] 00:30:18.110 }' 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:18.110 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:18.677 [2024-10-01 20:29:13.807143] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:30:18.677 [2024-10-01 20:29:13.807344] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:18.677 [2024-10-01 20:29:13.815160] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:30:18.677 [2024-10-01 20:29:13.818233] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:30:18.677 [2024-10-01 20:29:13.818493] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:30:18.677 [2024-10-01 20:29:13.818651] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:30:18.677 [2024-10-01 20:29:13.818694] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:30:18.677 [2024-10-01 20:29:13.818726] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:30:18.677 [2024-10-01 20:29:13.818752] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:18.677 "name": "Existed_Raid", 00:30:18.677 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:18.677 "strip_size_kb": 64, 00:30:18.677 "state": "configuring", 00:30:18.677 "raid_level": "raid5f", 00:30:18.677 "superblock": false, 00:30:18.677 "num_base_bdevs": 4, 00:30:18.677 "num_base_bdevs_discovered": 1, 00:30:18.677 "num_base_bdevs_operational": 4, 00:30:18.677 "base_bdevs_list": [ 00:30:18.677 { 00:30:18.677 "name": "BaseBdev1", 00:30:18.677 "uuid": "cb3a2648-6368-4b77-8e51-355764581eb3", 00:30:18.677 "is_configured": true, 00:30:18.677 "data_offset": 0, 00:30:18.677 "data_size": 65536 00:30:18.677 }, 00:30:18.677 { 00:30:18.677 "name": "BaseBdev2", 00:30:18.677 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:18.677 "is_configured": false, 00:30:18.677 "data_offset": 0, 00:30:18.677 "data_size": 0 00:30:18.677 }, 00:30:18.677 { 00:30:18.677 "name": "BaseBdev3", 00:30:18.677 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:18.677 "is_configured": false, 00:30:18.677 "data_offset": 0, 00:30:18.677 "data_size": 0 00:30:18.677 }, 00:30:18.677 { 00:30:18.677 "name": "BaseBdev4", 00:30:18.677 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:18.677 "is_configured": false, 00:30:18.677 "data_offset": 0, 00:30:18.677 "data_size": 0 00:30:18.677 } 00:30:18.677 ] 00:30:18.677 }' 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:18.677 20:29:13 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:19.244 [2024-10-01 20:29:14.379499] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:19.244 BaseBdev2 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:19.244 [ 00:30:19.244 { 00:30:19.244 "name": "BaseBdev2", 00:30:19.244 "aliases": [ 00:30:19.244 "9e2b291c-ab69-4f0c-8986-f39424c41f2a" 00:30:19.244 ], 00:30:19.244 "product_name": "Malloc disk", 00:30:19.244 "block_size": 512, 00:30:19.244 "num_blocks": 65536, 00:30:19.244 "uuid": "9e2b291c-ab69-4f0c-8986-f39424c41f2a", 00:30:19.244 "assigned_rate_limits": { 00:30:19.244 "rw_ios_per_sec": 0, 00:30:19.244 "rw_mbytes_per_sec": 0, 00:30:19.244 "r_mbytes_per_sec": 0, 00:30:19.244 "w_mbytes_per_sec": 0 00:30:19.244 }, 00:30:19.244 "claimed": true, 00:30:19.244 "claim_type": "exclusive_write", 00:30:19.244 "zoned": false, 00:30:19.244 "supported_io_types": { 00:30:19.244 "read": true, 00:30:19.244 "write": true, 00:30:19.244 "unmap": true, 00:30:19.244 "flush": true, 00:30:19.244 "reset": true, 00:30:19.244 "nvme_admin": false, 00:30:19.244 "nvme_io": false, 00:30:19.244 "nvme_io_md": false, 00:30:19.244 "write_zeroes": true, 00:30:19.244 "zcopy": true, 00:30:19.244 "get_zone_info": false, 00:30:19.244 "zone_management": false, 00:30:19.244 "zone_append": false, 00:30:19.244 "compare": false, 00:30:19.244 "compare_and_write": false, 00:30:19.244 "abort": true, 00:30:19.244 "seek_hole": false, 00:30:19.244 "seek_data": false, 00:30:19.244 "copy": true, 00:30:19.244 "nvme_iov_md": false 00:30:19.244 }, 00:30:19.244 "memory_domains": [ 00:30:19.244 { 00:30:19.244 "dma_device_id": "system", 00:30:19.244 "dma_device_type": 1 00:30:19.244 }, 00:30:19.244 { 00:30:19.244 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:19.244 "dma_device_type": 2 00:30:19.244 } 00:30:19.244 ], 00:30:19.244 "driver_specific": {} 00:30:19.244 } 00:30:19.244 ] 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:19.244 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:19.245 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:19.245 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:19.245 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:19.245 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:19.245 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:19.245 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:19.245 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:19.245 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:19.245 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:19.245 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:19.245 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:19.245 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:19.245 "name": "Existed_Raid", 00:30:19.245 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:19.245 "strip_size_kb": 64, 00:30:19.245 "state": "configuring", 00:30:19.245 "raid_level": "raid5f", 00:30:19.245 "superblock": false, 00:30:19.245 "num_base_bdevs": 4, 00:30:19.245 "num_base_bdevs_discovered": 2, 00:30:19.245 "num_base_bdevs_operational": 4, 00:30:19.245 "base_bdevs_list": [ 00:30:19.245 { 00:30:19.245 "name": "BaseBdev1", 00:30:19.245 "uuid": "cb3a2648-6368-4b77-8e51-355764581eb3", 00:30:19.245 "is_configured": true, 00:30:19.245 "data_offset": 0, 00:30:19.245 "data_size": 65536 00:30:19.245 }, 00:30:19.245 { 00:30:19.245 "name": "BaseBdev2", 00:30:19.245 "uuid": "9e2b291c-ab69-4f0c-8986-f39424c41f2a", 00:30:19.245 "is_configured": true, 00:30:19.245 "data_offset": 0, 00:30:19.245 "data_size": 65536 00:30:19.245 }, 00:30:19.245 { 00:30:19.245 "name": "BaseBdev3", 00:30:19.245 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:19.245 "is_configured": false, 00:30:19.245 "data_offset": 0, 00:30:19.245 "data_size": 0 00:30:19.245 }, 00:30:19.245 { 00:30:19.245 "name": "BaseBdev4", 00:30:19.245 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:19.245 "is_configured": false, 00:30:19.245 "data_offset": 0, 00:30:19.245 "data_size": 0 00:30:19.245 } 00:30:19.245 ] 00:30:19.245 }' 00:30:19.245 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:19.245 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:19.811 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:30:19.811 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:19.811 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:19.811 [2024-10-01 20:29:14.992264] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:30:19.811 BaseBdev3 00:30:19.811 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:19.811 20:29:14 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:30:19.811 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:30:19.811 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:19.811 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:30:19.811 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:19.811 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:19.811 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:19.811 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:19.811 20:29:14 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:19.811 [ 00:30:19.811 { 00:30:19.811 "name": "BaseBdev3", 00:30:19.811 "aliases": [ 00:30:19.811 "d1a07164-b3bd-4ab1-aad7-add698c01362" 00:30:19.811 ], 00:30:19.811 "product_name": "Malloc disk", 00:30:19.811 "block_size": 512, 00:30:19.811 "num_blocks": 65536, 00:30:19.811 "uuid": "d1a07164-b3bd-4ab1-aad7-add698c01362", 00:30:19.811 "assigned_rate_limits": { 00:30:19.811 "rw_ios_per_sec": 0, 00:30:19.811 "rw_mbytes_per_sec": 0, 00:30:19.811 "r_mbytes_per_sec": 0, 00:30:19.811 "w_mbytes_per_sec": 0 00:30:19.811 }, 00:30:19.811 "claimed": true, 00:30:19.811 "claim_type": "exclusive_write", 00:30:19.811 "zoned": false, 00:30:19.811 "supported_io_types": { 00:30:19.811 "read": true, 00:30:19.811 "write": true, 00:30:19.811 "unmap": true, 00:30:19.811 "flush": true, 00:30:19.811 "reset": true, 00:30:19.811 "nvme_admin": false, 00:30:19.811 "nvme_io": false, 00:30:19.811 "nvme_io_md": false, 00:30:19.811 "write_zeroes": true, 00:30:19.811 "zcopy": true, 00:30:19.811 "get_zone_info": false, 00:30:19.811 "zone_management": false, 00:30:19.811 "zone_append": false, 00:30:19.811 "compare": false, 00:30:19.811 "compare_and_write": false, 00:30:19.811 "abort": true, 00:30:19.811 "seek_hole": false, 00:30:19.811 "seek_data": false, 00:30:19.811 "copy": true, 00:30:19.811 "nvme_iov_md": false 00:30:19.811 }, 00:30:19.811 "memory_domains": [ 00:30:19.811 { 00:30:19.811 "dma_device_id": "system", 00:30:19.811 "dma_device_type": 1 00:30:19.811 }, 00:30:19.811 { 00:30:19.811 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:19.811 "dma_device_type": 2 00:30:19.811 } 00:30:19.811 ], 00:30:19.811 "driver_specific": {} 00:30:19.811 } 00:30:19.811 ] 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:19.811 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:20.069 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:20.069 "name": "Existed_Raid", 00:30:20.069 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:20.069 "strip_size_kb": 64, 00:30:20.069 "state": "configuring", 00:30:20.069 "raid_level": "raid5f", 00:30:20.069 "superblock": false, 00:30:20.069 "num_base_bdevs": 4, 00:30:20.069 "num_base_bdevs_discovered": 3, 00:30:20.069 "num_base_bdevs_operational": 4, 00:30:20.069 "base_bdevs_list": [ 00:30:20.069 { 00:30:20.069 "name": "BaseBdev1", 00:30:20.069 "uuid": "cb3a2648-6368-4b77-8e51-355764581eb3", 00:30:20.069 "is_configured": true, 00:30:20.069 "data_offset": 0, 00:30:20.069 "data_size": 65536 00:30:20.069 }, 00:30:20.069 { 00:30:20.069 "name": "BaseBdev2", 00:30:20.069 "uuid": "9e2b291c-ab69-4f0c-8986-f39424c41f2a", 00:30:20.069 "is_configured": true, 00:30:20.069 "data_offset": 0, 00:30:20.069 "data_size": 65536 00:30:20.069 }, 00:30:20.069 { 00:30:20.069 "name": "BaseBdev3", 00:30:20.069 "uuid": "d1a07164-b3bd-4ab1-aad7-add698c01362", 00:30:20.069 "is_configured": true, 00:30:20.069 "data_offset": 0, 00:30:20.069 "data_size": 65536 00:30:20.069 }, 00:30:20.069 { 00:30:20.069 "name": "BaseBdev4", 00:30:20.069 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:20.069 "is_configured": false, 00:30:20.069 "data_offset": 0, 00:30:20.069 "data_size": 0 00:30:20.069 } 00:30:20.069 ] 00:30:20.069 }' 00:30:20.069 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:20.069 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:20.327 [2024-10-01 20:29:15.535215] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:30:20.327 [2024-10-01 20:29:15.535573] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:30:20.327 [2024-10-01 20:29:15.535636] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:30:20.327 [2024-10-01 20:29:15.536124] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:30:20.327 [2024-10-01 20:29:15.542811] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:30:20.327 [2024-10-01 20:29:15.543003] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:30:20.327 [2024-10-01 20:29:15.543680] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:20.327 BaseBdev4 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:20.327 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:20.327 [ 00:30:20.328 { 00:30:20.328 "name": "BaseBdev4", 00:30:20.328 "aliases": [ 00:30:20.328 "2bba649c-1d39-497b-8853-b03f1ea24f48" 00:30:20.328 ], 00:30:20.328 "product_name": "Malloc disk", 00:30:20.328 "block_size": 512, 00:30:20.328 "num_blocks": 65536, 00:30:20.328 "uuid": "2bba649c-1d39-497b-8853-b03f1ea24f48", 00:30:20.328 "assigned_rate_limits": { 00:30:20.328 "rw_ios_per_sec": 0, 00:30:20.328 "rw_mbytes_per_sec": 0, 00:30:20.328 "r_mbytes_per_sec": 0, 00:30:20.328 "w_mbytes_per_sec": 0 00:30:20.328 }, 00:30:20.328 "claimed": true, 00:30:20.328 "claim_type": "exclusive_write", 00:30:20.328 "zoned": false, 00:30:20.328 "supported_io_types": { 00:30:20.328 "read": true, 00:30:20.328 "write": true, 00:30:20.328 "unmap": true, 00:30:20.328 "flush": true, 00:30:20.328 "reset": true, 00:30:20.328 "nvme_admin": false, 00:30:20.328 "nvme_io": false, 00:30:20.328 "nvme_io_md": false, 00:30:20.328 "write_zeroes": true, 00:30:20.328 "zcopy": true, 00:30:20.328 "get_zone_info": false, 00:30:20.328 "zone_management": false, 00:30:20.328 "zone_append": false, 00:30:20.328 "compare": false, 00:30:20.328 "compare_and_write": false, 00:30:20.328 "abort": true, 00:30:20.328 "seek_hole": false, 00:30:20.328 "seek_data": false, 00:30:20.328 "copy": true, 00:30:20.328 "nvme_iov_md": false 00:30:20.328 }, 00:30:20.328 "memory_domains": [ 00:30:20.328 { 00:30:20.328 "dma_device_id": "system", 00:30:20.328 "dma_device_type": 1 00:30:20.328 }, 00:30:20.328 { 00:30:20.328 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:20.328 "dma_device_type": 2 00:30:20.328 } 00:30:20.328 ], 00:30:20.328 "driver_specific": {} 00:30:20.328 } 00:30:20.328 ] 00:30:20.328 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:20.328 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:30:20.328 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:30:20.328 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:30:20.328 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:30:20.328 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:20.328 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:20.328 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:20.328 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:20.328 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:20.328 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:20.328 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:20.328 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:20.328 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:20.586 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:20.586 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:20.586 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:20.586 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:20.586 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:20.586 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:20.586 "name": "Existed_Raid", 00:30:20.586 "uuid": "01138f27-6ccb-479b-9f46-8ab61de4da51", 00:30:20.586 "strip_size_kb": 64, 00:30:20.586 "state": "online", 00:30:20.586 "raid_level": "raid5f", 00:30:20.586 "superblock": false, 00:30:20.586 "num_base_bdevs": 4, 00:30:20.586 "num_base_bdevs_discovered": 4, 00:30:20.586 "num_base_bdevs_operational": 4, 00:30:20.586 "base_bdevs_list": [ 00:30:20.586 { 00:30:20.586 "name": "BaseBdev1", 00:30:20.586 "uuid": "cb3a2648-6368-4b77-8e51-355764581eb3", 00:30:20.586 "is_configured": true, 00:30:20.586 "data_offset": 0, 00:30:20.586 "data_size": 65536 00:30:20.586 }, 00:30:20.586 { 00:30:20.586 "name": "BaseBdev2", 00:30:20.586 "uuid": "9e2b291c-ab69-4f0c-8986-f39424c41f2a", 00:30:20.586 "is_configured": true, 00:30:20.586 "data_offset": 0, 00:30:20.586 "data_size": 65536 00:30:20.586 }, 00:30:20.586 { 00:30:20.586 "name": "BaseBdev3", 00:30:20.586 "uuid": "d1a07164-b3bd-4ab1-aad7-add698c01362", 00:30:20.586 "is_configured": true, 00:30:20.586 "data_offset": 0, 00:30:20.586 "data_size": 65536 00:30:20.586 }, 00:30:20.586 { 00:30:20.586 "name": "BaseBdev4", 00:30:20.586 "uuid": "2bba649c-1d39-497b-8853-b03f1ea24f48", 00:30:20.586 "is_configured": true, 00:30:20.586 "data_offset": 0, 00:30:20.586 "data_size": 65536 00:30:20.586 } 00:30:20.586 ] 00:30:20.586 }' 00:30:20.586 20:29:15 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:20.586 20:29:15 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.153 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:30:21.153 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:30:21.153 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:30:21.153 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:30:21.153 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:30:21.153 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:30:21.153 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:30:21.153 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:21.153 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.153 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:30:21.153 [2024-10-01 20:29:16.107841] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:21.153 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:21.153 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:30:21.153 "name": "Existed_Raid", 00:30:21.153 "aliases": [ 00:30:21.153 "01138f27-6ccb-479b-9f46-8ab61de4da51" 00:30:21.153 ], 00:30:21.153 "product_name": "Raid Volume", 00:30:21.153 "block_size": 512, 00:30:21.153 "num_blocks": 196608, 00:30:21.153 "uuid": "01138f27-6ccb-479b-9f46-8ab61de4da51", 00:30:21.153 "assigned_rate_limits": { 00:30:21.153 "rw_ios_per_sec": 0, 00:30:21.153 "rw_mbytes_per_sec": 0, 00:30:21.153 "r_mbytes_per_sec": 0, 00:30:21.153 "w_mbytes_per_sec": 0 00:30:21.153 }, 00:30:21.153 "claimed": false, 00:30:21.153 "zoned": false, 00:30:21.153 "supported_io_types": { 00:30:21.153 "read": true, 00:30:21.153 "write": true, 00:30:21.153 "unmap": false, 00:30:21.153 "flush": false, 00:30:21.153 "reset": true, 00:30:21.153 "nvme_admin": false, 00:30:21.153 "nvme_io": false, 00:30:21.153 "nvme_io_md": false, 00:30:21.153 "write_zeroes": true, 00:30:21.153 "zcopy": false, 00:30:21.153 "get_zone_info": false, 00:30:21.153 "zone_management": false, 00:30:21.153 "zone_append": false, 00:30:21.153 "compare": false, 00:30:21.153 "compare_and_write": false, 00:30:21.153 "abort": false, 00:30:21.153 "seek_hole": false, 00:30:21.153 "seek_data": false, 00:30:21.153 "copy": false, 00:30:21.153 "nvme_iov_md": false 00:30:21.153 }, 00:30:21.153 "driver_specific": { 00:30:21.153 "raid": { 00:30:21.153 "uuid": "01138f27-6ccb-479b-9f46-8ab61de4da51", 00:30:21.153 "strip_size_kb": 64, 00:30:21.153 "state": "online", 00:30:21.153 "raid_level": "raid5f", 00:30:21.153 "superblock": false, 00:30:21.153 "num_base_bdevs": 4, 00:30:21.153 "num_base_bdevs_discovered": 4, 00:30:21.153 "num_base_bdevs_operational": 4, 00:30:21.153 "base_bdevs_list": [ 00:30:21.153 { 00:30:21.153 "name": "BaseBdev1", 00:30:21.153 "uuid": "cb3a2648-6368-4b77-8e51-355764581eb3", 00:30:21.153 "is_configured": true, 00:30:21.153 "data_offset": 0, 00:30:21.153 "data_size": 65536 00:30:21.153 }, 00:30:21.153 { 00:30:21.153 "name": "BaseBdev2", 00:30:21.153 "uuid": "9e2b291c-ab69-4f0c-8986-f39424c41f2a", 00:30:21.153 "is_configured": true, 00:30:21.153 "data_offset": 0, 00:30:21.153 "data_size": 65536 00:30:21.153 }, 00:30:21.153 { 00:30:21.153 "name": "BaseBdev3", 00:30:21.154 "uuid": "d1a07164-b3bd-4ab1-aad7-add698c01362", 00:30:21.154 "is_configured": true, 00:30:21.154 "data_offset": 0, 00:30:21.154 "data_size": 65536 00:30:21.154 }, 00:30:21.154 { 00:30:21.154 "name": "BaseBdev4", 00:30:21.154 "uuid": "2bba649c-1d39-497b-8853-b03f1ea24f48", 00:30:21.154 "is_configured": true, 00:30:21.154 "data_offset": 0, 00:30:21.154 "data_size": 65536 00:30:21.154 } 00:30:21.154 ] 00:30:21.154 } 00:30:21.154 } 00:30:21.154 }' 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:30:21.154 BaseBdev2 00:30:21.154 BaseBdev3 00:30:21.154 BaseBdev4' 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:21.154 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.461 [2024-10-01 20:29:16.487663] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:21.461 "name": "Existed_Raid", 00:30:21.461 "uuid": "01138f27-6ccb-479b-9f46-8ab61de4da51", 00:30:21.461 "strip_size_kb": 64, 00:30:21.461 "state": "online", 00:30:21.461 "raid_level": "raid5f", 00:30:21.461 "superblock": false, 00:30:21.461 "num_base_bdevs": 4, 00:30:21.461 "num_base_bdevs_discovered": 3, 00:30:21.461 "num_base_bdevs_operational": 3, 00:30:21.461 "base_bdevs_list": [ 00:30:21.461 { 00:30:21.461 "name": null, 00:30:21.461 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:21.461 "is_configured": false, 00:30:21.461 "data_offset": 0, 00:30:21.461 "data_size": 65536 00:30:21.461 }, 00:30:21.461 { 00:30:21.461 "name": "BaseBdev2", 00:30:21.461 "uuid": "9e2b291c-ab69-4f0c-8986-f39424c41f2a", 00:30:21.461 "is_configured": true, 00:30:21.461 "data_offset": 0, 00:30:21.461 "data_size": 65536 00:30:21.461 }, 00:30:21.461 { 00:30:21.461 "name": "BaseBdev3", 00:30:21.461 "uuid": "d1a07164-b3bd-4ab1-aad7-add698c01362", 00:30:21.461 "is_configured": true, 00:30:21.461 "data_offset": 0, 00:30:21.461 "data_size": 65536 00:30:21.461 }, 00:30:21.461 { 00:30:21.461 "name": "BaseBdev4", 00:30:21.461 "uuid": "2bba649c-1d39-497b-8853-b03f1ea24f48", 00:30:21.461 "is_configured": true, 00:30:21.461 "data_offset": 0, 00:30:21.461 "data_size": 65536 00:30:21.461 } 00:30:21.461 ] 00:30:21.461 }' 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:21.461 20:29:16 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.028 [2024-10-01 20:29:17.166161] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:30:22.028 [2024-10-01 20:29:17.166493] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:22.028 [2024-10-01 20:29:17.242853] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:30:22.028 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.287 [2024-10-01 20:29:17.306976] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.287 [2024-10-01 20:29:17.454583] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:30:22.287 [2024-10-01 20:29:17.454865] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:22.287 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.547 BaseBdev2 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:22.547 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.548 [ 00:30:22.548 { 00:30:22.548 "name": "BaseBdev2", 00:30:22.548 "aliases": [ 00:30:22.548 "e294b34e-eee7-4477-95fc-b59c454dc224" 00:30:22.548 ], 00:30:22.548 "product_name": "Malloc disk", 00:30:22.548 "block_size": 512, 00:30:22.548 "num_blocks": 65536, 00:30:22.548 "uuid": "e294b34e-eee7-4477-95fc-b59c454dc224", 00:30:22.548 "assigned_rate_limits": { 00:30:22.548 "rw_ios_per_sec": 0, 00:30:22.548 "rw_mbytes_per_sec": 0, 00:30:22.548 "r_mbytes_per_sec": 0, 00:30:22.548 "w_mbytes_per_sec": 0 00:30:22.548 }, 00:30:22.548 "claimed": false, 00:30:22.548 "zoned": false, 00:30:22.548 "supported_io_types": { 00:30:22.548 "read": true, 00:30:22.548 "write": true, 00:30:22.548 "unmap": true, 00:30:22.548 "flush": true, 00:30:22.548 "reset": true, 00:30:22.548 "nvme_admin": false, 00:30:22.548 "nvme_io": false, 00:30:22.548 "nvme_io_md": false, 00:30:22.548 "write_zeroes": true, 00:30:22.548 "zcopy": true, 00:30:22.548 "get_zone_info": false, 00:30:22.548 "zone_management": false, 00:30:22.548 "zone_append": false, 00:30:22.548 "compare": false, 00:30:22.548 "compare_and_write": false, 00:30:22.548 "abort": true, 00:30:22.548 "seek_hole": false, 00:30:22.548 "seek_data": false, 00:30:22.548 "copy": true, 00:30:22.548 "nvme_iov_md": false 00:30:22.548 }, 00:30:22.548 "memory_domains": [ 00:30:22.548 { 00:30:22.548 "dma_device_id": "system", 00:30:22.548 "dma_device_type": 1 00:30:22.548 }, 00:30:22.548 { 00:30:22.548 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:22.548 "dma_device_type": 2 00:30:22.548 } 00:30:22.548 ], 00:30:22.548 "driver_specific": {} 00:30:22.548 } 00:30:22.548 ] 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.548 BaseBdev3 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.548 [ 00:30:22.548 { 00:30:22.548 "name": "BaseBdev3", 00:30:22.548 "aliases": [ 00:30:22.548 "e31bb439-0414-42b0-96ed-8319b3f03ad8" 00:30:22.548 ], 00:30:22.548 "product_name": "Malloc disk", 00:30:22.548 "block_size": 512, 00:30:22.548 "num_blocks": 65536, 00:30:22.548 "uuid": "e31bb439-0414-42b0-96ed-8319b3f03ad8", 00:30:22.548 "assigned_rate_limits": { 00:30:22.548 "rw_ios_per_sec": 0, 00:30:22.548 "rw_mbytes_per_sec": 0, 00:30:22.548 "r_mbytes_per_sec": 0, 00:30:22.548 "w_mbytes_per_sec": 0 00:30:22.548 }, 00:30:22.548 "claimed": false, 00:30:22.548 "zoned": false, 00:30:22.548 "supported_io_types": { 00:30:22.548 "read": true, 00:30:22.548 "write": true, 00:30:22.548 "unmap": true, 00:30:22.548 "flush": true, 00:30:22.548 "reset": true, 00:30:22.548 "nvme_admin": false, 00:30:22.548 "nvme_io": false, 00:30:22.548 "nvme_io_md": false, 00:30:22.548 "write_zeroes": true, 00:30:22.548 "zcopy": true, 00:30:22.548 "get_zone_info": false, 00:30:22.548 "zone_management": false, 00:30:22.548 "zone_append": false, 00:30:22.548 "compare": false, 00:30:22.548 "compare_and_write": false, 00:30:22.548 "abort": true, 00:30:22.548 "seek_hole": false, 00:30:22.548 "seek_data": false, 00:30:22.548 "copy": true, 00:30:22.548 "nvme_iov_md": false 00:30:22.548 }, 00:30:22.548 "memory_domains": [ 00:30:22.548 { 00:30:22.548 "dma_device_id": "system", 00:30:22.548 "dma_device_type": 1 00:30:22.548 }, 00:30:22.548 { 00:30:22.548 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:22.548 "dma_device_type": 2 00:30:22.548 } 00:30:22.548 ], 00:30:22.548 "driver_specific": {} 00:30:22.548 } 00:30:22.548 ] 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.548 BaseBdev4 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.548 [ 00:30:22.548 { 00:30:22.548 "name": "BaseBdev4", 00:30:22.548 "aliases": [ 00:30:22.548 "8f73e27d-3d29-43e9-bf90-5b8af17f7ed2" 00:30:22.548 ], 00:30:22.548 "product_name": "Malloc disk", 00:30:22.548 "block_size": 512, 00:30:22.548 "num_blocks": 65536, 00:30:22.548 "uuid": "8f73e27d-3d29-43e9-bf90-5b8af17f7ed2", 00:30:22.548 "assigned_rate_limits": { 00:30:22.548 "rw_ios_per_sec": 0, 00:30:22.548 "rw_mbytes_per_sec": 0, 00:30:22.548 "r_mbytes_per_sec": 0, 00:30:22.548 "w_mbytes_per_sec": 0 00:30:22.548 }, 00:30:22.548 "claimed": false, 00:30:22.548 "zoned": false, 00:30:22.548 "supported_io_types": { 00:30:22.548 "read": true, 00:30:22.548 "write": true, 00:30:22.548 "unmap": true, 00:30:22.548 "flush": true, 00:30:22.548 "reset": true, 00:30:22.548 "nvme_admin": false, 00:30:22.548 "nvme_io": false, 00:30:22.548 "nvme_io_md": false, 00:30:22.548 "write_zeroes": true, 00:30:22.548 "zcopy": true, 00:30:22.548 "get_zone_info": false, 00:30:22.548 "zone_management": false, 00:30:22.548 "zone_append": false, 00:30:22.548 "compare": false, 00:30:22.548 "compare_and_write": false, 00:30:22.548 "abort": true, 00:30:22.548 "seek_hole": false, 00:30:22.548 "seek_data": false, 00:30:22.548 "copy": true, 00:30:22.548 "nvme_iov_md": false 00:30:22.548 }, 00:30:22.548 "memory_domains": [ 00:30:22.548 { 00:30:22.548 "dma_device_id": "system", 00:30:22.548 "dma_device_type": 1 00:30:22.548 }, 00:30:22.548 { 00:30:22.548 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:22.548 "dma_device_type": 2 00:30:22.548 } 00:30:22.548 ], 00:30:22.548 "driver_specific": {} 00:30:22.548 } 00:30:22.548 ] 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:30:22.548 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:30:22.549 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:30:22.549 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:30:22.549 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.549 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.549 [2024-10-01 20:29:17.790840] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:30:22.549 [2024-10-01 20:29:17.791045] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:30:22.549 [2024-10-01 20:29:17.791212] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:22.549 [2024-10-01 20:29:17.793956] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:30:22.549 [2024-10-01 20:29:17.794204] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:30:22.549 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.549 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:22.549 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:22.549 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:22.549 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:22.549 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:22.549 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:22.549 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:22.549 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:22.549 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:22.549 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:22.808 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:22.808 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:22.808 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:22.808 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:22.808 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:22.808 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:22.808 "name": "Existed_Raid", 00:30:22.808 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:22.808 "strip_size_kb": 64, 00:30:22.808 "state": "configuring", 00:30:22.808 "raid_level": "raid5f", 00:30:22.808 "superblock": false, 00:30:22.808 "num_base_bdevs": 4, 00:30:22.808 "num_base_bdevs_discovered": 3, 00:30:22.808 "num_base_bdevs_operational": 4, 00:30:22.808 "base_bdevs_list": [ 00:30:22.808 { 00:30:22.808 "name": "BaseBdev1", 00:30:22.808 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:22.808 "is_configured": false, 00:30:22.808 "data_offset": 0, 00:30:22.808 "data_size": 0 00:30:22.808 }, 00:30:22.808 { 00:30:22.808 "name": "BaseBdev2", 00:30:22.808 "uuid": "e294b34e-eee7-4477-95fc-b59c454dc224", 00:30:22.808 "is_configured": true, 00:30:22.808 "data_offset": 0, 00:30:22.808 "data_size": 65536 00:30:22.808 }, 00:30:22.808 { 00:30:22.808 "name": "BaseBdev3", 00:30:22.808 "uuid": "e31bb439-0414-42b0-96ed-8319b3f03ad8", 00:30:22.808 "is_configured": true, 00:30:22.808 "data_offset": 0, 00:30:22.808 "data_size": 65536 00:30:22.808 }, 00:30:22.808 { 00:30:22.808 "name": "BaseBdev4", 00:30:22.808 "uuid": "8f73e27d-3d29-43e9-bf90-5b8af17f7ed2", 00:30:22.808 "is_configured": true, 00:30:22.808 "data_offset": 0, 00:30:22.808 "data_size": 65536 00:30:22.808 } 00:30:22.808 ] 00:30:22.808 }' 00:30:22.808 20:29:17 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:22.808 20:29:17 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:23.377 [2024-10-01 20:29:18.339052] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:23.377 "name": "Existed_Raid", 00:30:23.377 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:23.377 "strip_size_kb": 64, 00:30:23.377 "state": "configuring", 00:30:23.377 "raid_level": "raid5f", 00:30:23.377 "superblock": false, 00:30:23.377 "num_base_bdevs": 4, 00:30:23.377 "num_base_bdevs_discovered": 2, 00:30:23.377 "num_base_bdevs_operational": 4, 00:30:23.377 "base_bdevs_list": [ 00:30:23.377 { 00:30:23.377 "name": "BaseBdev1", 00:30:23.377 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:23.377 "is_configured": false, 00:30:23.377 "data_offset": 0, 00:30:23.377 "data_size": 0 00:30:23.377 }, 00:30:23.377 { 00:30:23.377 "name": null, 00:30:23.377 "uuid": "e294b34e-eee7-4477-95fc-b59c454dc224", 00:30:23.377 "is_configured": false, 00:30:23.377 "data_offset": 0, 00:30:23.377 "data_size": 65536 00:30:23.377 }, 00:30:23.377 { 00:30:23.377 "name": "BaseBdev3", 00:30:23.377 "uuid": "e31bb439-0414-42b0-96ed-8319b3f03ad8", 00:30:23.377 "is_configured": true, 00:30:23.377 "data_offset": 0, 00:30:23.377 "data_size": 65536 00:30:23.377 }, 00:30:23.377 { 00:30:23.377 "name": "BaseBdev4", 00:30:23.377 "uuid": "8f73e27d-3d29-43e9-bf90-5b8af17f7ed2", 00:30:23.377 "is_configured": true, 00:30:23.377 "data_offset": 0, 00:30:23.377 "data_size": 65536 00:30:23.377 } 00:30:23.377 ] 00:30:23.377 }' 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:23.377 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:23.946 [2024-10-01 20:29:18.979609] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:30:23.946 BaseBdev1 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:23.946 20:29:18 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:23.946 [ 00:30:23.946 { 00:30:23.946 "name": "BaseBdev1", 00:30:23.946 "aliases": [ 00:30:23.946 "9f9bdb8c-da17-4dd7-b75a-b48572e70654" 00:30:23.946 ], 00:30:23.946 "product_name": "Malloc disk", 00:30:23.946 "block_size": 512, 00:30:23.946 "num_blocks": 65536, 00:30:23.946 "uuid": "9f9bdb8c-da17-4dd7-b75a-b48572e70654", 00:30:23.946 "assigned_rate_limits": { 00:30:23.946 "rw_ios_per_sec": 0, 00:30:23.946 "rw_mbytes_per_sec": 0, 00:30:23.946 "r_mbytes_per_sec": 0, 00:30:23.946 "w_mbytes_per_sec": 0 00:30:23.946 }, 00:30:23.946 "claimed": true, 00:30:23.946 "claim_type": "exclusive_write", 00:30:23.946 "zoned": false, 00:30:23.946 "supported_io_types": { 00:30:23.946 "read": true, 00:30:23.946 "write": true, 00:30:23.946 "unmap": true, 00:30:23.946 "flush": true, 00:30:23.946 "reset": true, 00:30:23.946 "nvme_admin": false, 00:30:23.946 "nvme_io": false, 00:30:23.946 "nvme_io_md": false, 00:30:23.946 "write_zeroes": true, 00:30:23.946 "zcopy": true, 00:30:23.946 "get_zone_info": false, 00:30:23.946 "zone_management": false, 00:30:23.946 "zone_append": false, 00:30:23.946 "compare": false, 00:30:23.946 "compare_and_write": false, 00:30:23.946 "abort": true, 00:30:23.946 "seek_hole": false, 00:30:23.946 "seek_data": false, 00:30:23.946 "copy": true, 00:30:23.946 "nvme_iov_md": false 00:30:23.946 }, 00:30:23.946 "memory_domains": [ 00:30:23.946 { 00:30:23.946 "dma_device_id": "system", 00:30:23.946 "dma_device_type": 1 00:30:23.946 }, 00:30:23.946 { 00:30:23.946 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:23.946 "dma_device_type": 2 00:30:23.946 } 00:30:23.946 ], 00:30:23.946 "driver_specific": {} 00:30:23.946 } 00:30:23.946 ] 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:23.946 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:23.946 "name": "Existed_Raid", 00:30:23.946 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:23.946 "strip_size_kb": 64, 00:30:23.946 "state": "configuring", 00:30:23.946 "raid_level": "raid5f", 00:30:23.946 "superblock": false, 00:30:23.946 "num_base_bdevs": 4, 00:30:23.946 "num_base_bdevs_discovered": 3, 00:30:23.947 "num_base_bdevs_operational": 4, 00:30:23.947 "base_bdevs_list": [ 00:30:23.947 { 00:30:23.947 "name": "BaseBdev1", 00:30:23.947 "uuid": "9f9bdb8c-da17-4dd7-b75a-b48572e70654", 00:30:23.947 "is_configured": true, 00:30:23.947 "data_offset": 0, 00:30:23.947 "data_size": 65536 00:30:23.947 }, 00:30:23.947 { 00:30:23.947 "name": null, 00:30:23.947 "uuid": "e294b34e-eee7-4477-95fc-b59c454dc224", 00:30:23.947 "is_configured": false, 00:30:23.947 "data_offset": 0, 00:30:23.947 "data_size": 65536 00:30:23.947 }, 00:30:23.947 { 00:30:23.947 "name": "BaseBdev3", 00:30:23.947 "uuid": "e31bb439-0414-42b0-96ed-8319b3f03ad8", 00:30:23.947 "is_configured": true, 00:30:23.947 "data_offset": 0, 00:30:23.947 "data_size": 65536 00:30:23.947 }, 00:30:23.947 { 00:30:23.947 "name": "BaseBdev4", 00:30:23.947 "uuid": "8f73e27d-3d29-43e9-bf90-5b8af17f7ed2", 00:30:23.947 "is_configured": true, 00:30:23.947 "data_offset": 0, 00:30:23.947 "data_size": 65536 00:30:23.947 } 00:30:23.947 ] 00:30:23.947 }' 00:30:23.947 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:23.947 20:29:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:24.514 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:24.514 20:29:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:24.514 20:29:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:24.514 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:30:24.514 20:29:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:24.514 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:30:24.514 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:24.515 [2024-10-01 20:29:19.591890] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:24.515 "name": "Existed_Raid", 00:30:24.515 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:24.515 "strip_size_kb": 64, 00:30:24.515 "state": "configuring", 00:30:24.515 "raid_level": "raid5f", 00:30:24.515 "superblock": false, 00:30:24.515 "num_base_bdevs": 4, 00:30:24.515 "num_base_bdevs_discovered": 2, 00:30:24.515 "num_base_bdevs_operational": 4, 00:30:24.515 "base_bdevs_list": [ 00:30:24.515 { 00:30:24.515 "name": "BaseBdev1", 00:30:24.515 "uuid": "9f9bdb8c-da17-4dd7-b75a-b48572e70654", 00:30:24.515 "is_configured": true, 00:30:24.515 "data_offset": 0, 00:30:24.515 "data_size": 65536 00:30:24.515 }, 00:30:24.515 { 00:30:24.515 "name": null, 00:30:24.515 "uuid": "e294b34e-eee7-4477-95fc-b59c454dc224", 00:30:24.515 "is_configured": false, 00:30:24.515 "data_offset": 0, 00:30:24.515 "data_size": 65536 00:30:24.515 }, 00:30:24.515 { 00:30:24.515 "name": null, 00:30:24.515 "uuid": "e31bb439-0414-42b0-96ed-8319b3f03ad8", 00:30:24.515 "is_configured": false, 00:30:24.515 "data_offset": 0, 00:30:24.515 "data_size": 65536 00:30:24.515 }, 00:30:24.515 { 00:30:24.515 "name": "BaseBdev4", 00:30:24.515 "uuid": "8f73e27d-3d29-43e9-bf90-5b8af17f7ed2", 00:30:24.515 "is_configured": true, 00:30:24.515 "data_offset": 0, 00:30:24.515 "data_size": 65536 00:30:24.515 } 00:30:24.515 ] 00:30:24.515 }' 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:24.515 20:29:19 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:25.085 [2024-10-01 20:29:20.168160] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:25.085 "name": "Existed_Raid", 00:30:25.085 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:25.085 "strip_size_kb": 64, 00:30:25.085 "state": "configuring", 00:30:25.085 "raid_level": "raid5f", 00:30:25.085 "superblock": false, 00:30:25.085 "num_base_bdevs": 4, 00:30:25.085 "num_base_bdevs_discovered": 3, 00:30:25.085 "num_base_bdevs_operational": 4, 00:30:25.085 "base_bdevs_list": [ 00:30:25.085 { 00:30:25.085 "name": "BaseBdev1", 00:30:25.085 "uuid": "9f9bdb8c-da17-4dd7-b75a-b48572e70654", 00:30:25.085 "is_configured": true, 00:30:25.085 "data_offset": 0, 00:30:25.085 "data_size": 65536 00:30:25.085 }, 00:30:25.085 { 00:30:25.085 "name": null, 00:30:25.085 "uuid": "e294b34e-eee7-4477-95fc-b59c454dc224", 00:30:25.085 "is_configured": false, 00:30:25.085 "data_offset": 0, 00:30:25.085 "data_size": 65536 00:30:25.085 }, 00:30:25.085 { 00:30:25.085 "name": "BaseBdev3", 00:30:25.085 "uuid": "e31bb439-0414-42b0-96ed-8319b3f03ad8", 00:30:25.085 "is_configured": true, 00:30:25.085 "data_offset": 0, 00:30:25.085 "data_size": 65536 00:30:25.085 }, 00:30:25.085 { 00:30:25.085 "name": "BaseBdev4", 00:30:25.085 "uuid": "8f73e27d-3d29-43e9-bf90-5b8af17f7ed2", 00:30:25.085 "is_configured": true, 00:30:25.085 "data_offset": 0, 00:30:25.085 "data_size": 65536 00:30:25.085 } 00:30:25.085 ] 00:30:25.085 }' 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:25.085 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:25.653 [2024-10-01 20:29:20.732289] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:25.653 "name": "Existed_Raid", 00:30:25.653 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:25.653 "strip_size_kb": 64, 00:30:25.653 "state": "configuring", 00:30:25.653 "raid_level": "raid5f", 00:30:25.653 "superblock": false, 00:30:25.653 "num_base_bdevs": 4, 00:30:25.653 "num_base_bdevs_discovered": 2, 00:30:25.653 "num_base_bdevs_operational": 4, 00:30:25.653 "base_bdevs_list": [ 00:30:25.653 { 00:30:25.653 "name": null, 00:30:25.653 "uuid": "9f9bdb8c-da17-4dd7-b75a-b48572e70654", 00:30:25.653 "is_configured": false, 00:30:25.653 "data_offset": 0, 00:30:25.653 "data_size": 65536 00:30:25.653 }, 00:30:25.653 { 00:30:25.653 "name": null, 00:30:25.653 "uuid": "e294b34e-eee7-4477-95fc-b59c454dc224", 00:30:25.653 "is_configured": false, 00:30:25.653 "data_offset": 0, 00:30:25.653 "data_size": 65536 00:30:25.653 }, 00:30:25.653 { 00:30:25.653 "name": "BaseBdev3", 00:30:25.653 "uuid": "e31bb439-0414-42b0-96ed-8319b3f03ad8", 00:30:25.653 "is_configured": true, 00:30:25.653 "data_offset": 0, 00:30:25.653 "data_size": 65536 00:30:25.653 }, 00:30:25.653 { 00:30:25.653 "name": "BaseBdev4", 00:30:25.653 "uuid": "8f73e27d-3d29-43e9-bf90-5b8af17f7ed2", 00:30:25.653 "is_configured": true, 00:30:25.653 "data_offset": 0, 00:30:25.653 "data_size": 65536 00:30:25.653 } 00:30:25.653 ] 00:30:25.653 }' 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:25.653 20:29:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:26.221 [2024-10-01 20:29:21.386931] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:26.221 "name": "Existed_Raid", 00:30:26.221 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:26.221 "strip_size_kb": 64, 00:30:26.221 "state": "configuring", 00:30:26.221 "raid_level": "raid5f", 00:30:26.221 "superblock": false, 00:30:26.221 "num_base_bdevs": 4, 00:30:26.221 "num_base_bdevs_discovered": 3, 00:30:26.221 "num_base_bdevs_operational": 4, 00:30:26.221 "base_bdevs_list": [ 00:30:26.221 { 00:30:26.221 "name": null, 00:30:26.221 "uuid": "9f9bdb8c-da17-4dd7-b75a-b48572e70654", 00:30:26.221 "is_configured": false, 00:30:26.221 "data_offset": 0, 00:30:26.221 "data_size": 65536 00:30:26.221 }, 00:30:26.221 { 00:30:26.221 "name": "BaseBdev2", 00:30:26.221 "uuid": "e294b34e-eee7-4477-95fc-b59c454dc224", 00:30:26.221 "is_configured": true, 00:30:26.221 "data_offset": 0, 00:30:26.221 "data_size": 65536 00:30:26.221 }, 00:30:26.221 { 00:30:26.221 "name": "BaseBdev3", 00:30:26.221 "uuid": "e31bb439-0414-42b0-96ed-8319b3f03ad8", 00:30:26.221 "is_configured": true, 00:30:26.221 "data_offset": 0, 00:30:26.221 "data_size": 65536 00:30:26.221 }, 00:30:26.221 { 00:30:26.221 "name": "BaseBdev4", 00:30:26.221 "uuid": "8f73e27d-3d29-43e9-bf90-5b8af17f7ed2", 00:30:26.221 "is_configured": true, 00:30:26.221 "data_offset": 0, 00:30:26.221 "data_size": 65536 00:30:26.221 } 00:30:26.221 ] 00:30:26.221 }' 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:26.221 20:29:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:26.789 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:26.789 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:30:26.789 20:29:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:26.789 20:29:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:26.789 20:29:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:26.789 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:30:26.789 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:26.789 20:29:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:26.789 20:29:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:26.789 20:29:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:30:26.789 20:29:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:26.789 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 9f9bdb8c-da17-4dd7-b75a-b48572e70654 00:30:26.789 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:26.789 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:27.048 [2024-10-01 20:29:22.084238] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:30:27.048 [2024-10-01 20:29:22.084536] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:30:27.048 [2024-10-01 20:29:22.084564] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:30:27.048 [2024-10-01 20:29:22.084947] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:30:27.048 [2024-10-01 20:29:22.093086] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:30:27.048 [2024-10-01 20:29:22.093289] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:30:27.048 [2024-10-01 20:29:22.093822] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:27.048 NewBaseBdev 00:30:27.048 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:27.048 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:30:27.048 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:30:27.048 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:27.048 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:30:27.048 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:27.048 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:27.048 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:27.048 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:27.048 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:27.048 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:27.048 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:30:27.048 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:27.048 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:27.048 [ 00:30:27.048 { 00:30:27.048 "name": "NewBaseBdev", 00:30:27.048 "aliases": [ 00:30:27.049 "9f9bdb8c-da17-4dd7-b75a-b48572e70654" 00:30:27.049 ], 00:30:27.049 "product_name": "Malloc disk", 00:30:27.049 "block_size": 512, 00:30:27.049 "num_blocks": 65536, 00:30:27.049 "uuid": "9f9bdb8c-da17-4dd7-b75a-b48572e70654", 00:30:27.049 "assigned_rate_limits": { 00:30:27.049 "rw_ios_per_sec": 0, 00:30:27.049 "rw_mbytes_per_sec": 0, 00:30:27.049 "r_mbytes_per_sec": 0, 00:30:27.049 "w_mbytes_per_sec": 0 00:30:27.049 }, 00:30:27.049 "claimed": true, 00:30:27.049 "claim_type": "exclusive_write", 00:30:27.049 "zoned": false, 00:30:27.049 "supported_io_types": { 00:30:27.049 "read": true, 00:30:27.049 "write": true, 00:30:27.049 "unmap": true, 00:30:27.049 "flush": true, 00:30:27.049 "reset": true, 00:30:27.049 "nvme_admin": false, 00:30:27.049 "nvme_io": false, 00:30:27.049 "nvme_io_md": false, 00:30:27.049 "write_zeroes": true, 00:30:27.049 "zcopy": true, 00:30:27.049 "get_zone_info": false, 00:30:27.049 "zone_management": false, 00:30:27.049 "zone_append": false, 00:30:27.049 "compare": false, 00:30:27.049 "compare_and_write": false, 00:30:27.049 "abort": true, 00:30:27.049 "seek_hole": false, 00:30:27.049 "seek_data": false, 00:30:27.049 "copy": true, 00:30:27.049 "nvme_iov_md": false 00:30:27.049 }, 00:30:27.049 "memory_domains": [ 00:30:27.049 { 00:30:27.049 "dma_device_id": "system", 00:30:27.049 "dma_device_type": 1 00:30:27.049 }, 00:30:27.049 { 00:30:27.049 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:27.049 "dma_device_type": 2 00:30:27.049 } 00:30:27.049 ], 00:30:27.049 "driver_specific": {} 00:30:27.049 } 00:30:27.049 ] 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:27.049 "name": "Existed_Raid", 00:30:27.049 "uuid": "125d1438-35ed-4338-9b9f-543e53a3ddd7", 00:30:27.049 "strip_size_kb": 64, 00:30:27.049 "state": "online", 00:30:27.049 "raid_level": "raid5f", 00:30:27.049 "superblock": false, 00:30:27.049 "num_base_bdevs": 4, 00:30:27.049 "num_base_bdevs_discovered": 4, 00:30:27.049 "num_base_bdevs_operational": 4, 00:30:27.049 "base_bdevs_list": [ 00:30:27.049 { 00:30:27.049 "name": "NewBaseBdev", 00:30:27.049 "uuid": "9f9bdb8c-da17-4dd7-b75a-b48572e70654", 00:30:27.049 "is_configured": true, 00:30:27.049 "data_offset": 0, 00:30:27.049 "data_size": 65536 00:30:27.049 }, 00:30:27.049 { 00:30:27.049 "name": "BaseBdev2", 00:30:27.049 "uuid": "e294b34e-eee7-4477-95fc-b59c454dc224", 00:30:27.049 "is_configured": true, 00:30:27.049 "data_offset": 0, 00:30:27.049 "data_size": 65536 00:30:27.049 }, 00:30:27.049 { 00:30:27.049 "name": "BaseBdev3", 00:30:27.049 "uuid": "e31bb439-0414-42b0-96ed-8319b3f03ad8", 00:30:27.049 "is_configured": true, 00:30:27.049 "data_offset": 0, 00:30:27.049 "data_size": 65536 00:30:27.049 }, 00:30:27.049 { 00:30:27.049 "name": "BaseBdev4", 00:30:27.049 "uuid": "8f73e27d-3d29-43e9-bf90-5b8af17f7ed2", 00:30:27.049 "is_configured": true, 00:30:27.049 "data_offset": 0, 00:30:27.049 "data_size": 65536 00:30:27.049 } 00:30:27.049 ] 00:30:27.049 }' 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:27.049 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:27.617 [2024-10-01 20:29:22.678001] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:30:27.617 "name": "Existed_Raid", 00:30:27.617 "aliases": [ 00:30:27.617 "125d1438-35ed-4338-9b9f-543e53a3ddd7" 00:30:27.617 ], 00:30:27.617 "product_name": "Raid Volume", 00:30:27.617 "block_size": 512, 00:30:27.617 "num_blocks": 196608, 00:30:27.617 "uuid": "125d1438-35ed-4338-9b9f-543e53a3ddd7", 00:30:27.617 "assigned_rate_limits": { 00:30:27.617 "rw_ios_per_sec": 0, 00:30:27.617 "rw_mbytes_per_sec": 0, 00:30:27.617 "r_mbytes_per_sec": 0, 00:30:27.617 "w_mbytes_per_sec": 0 00:30:27.617 }, 00:30:27.617 "claimed": false, 00:30:27.617 "zoned": false, 00:30:27.617 "supported_io_types": { 00:30:27.617 "read": true, 00:30:27.617 "write": true, 00:30:27.617 "unmap": false, 00:30:27.617 "flush": false, 00:30:27.617 "reset": true, 00:30:27.617 "nvme_admin": false, 00:30:27.617 "nvme_io": false, 00:30:27.617 "nvme_io_md": false, 00:30:27.617 "write_zeroes": true, 00:30:27.617 "zcopy": false, 00:30:27.617 "get_zone_info": false, 00:30:27.617 "zone_management": false, 00:30:27.617 "zone_append": false, 00:30:27.617 "compare": false, 00:30:27.617 "compare_and_write": false, 00:30:27.617 "abort": false, 00:30:27.617 "seek_hole": false, 00:30:27.617 "seek_data": false, 00:30:27.617 "copy": false, 00:30:27.617 "nvme_iov_md": false 00:30:27.617 }, 00:30:27.617 "driver_specific": { 00:30:27.617 "raid": { 00:30:27.617 "uuid": "125d1438-35ed-4338-9b9f-543e53a3ddd7", 00:30:27.617 "strip_size_kb": 64, 00:30:27.617 "state": "online", 00:30:27.617 "raid_level": "raid5f", 00:30:27.617 "superblock": false, 00:30:27.617 "num_base_bdevs": 4, 00:30:27.617 "num_base_bdevs_discovered": 4, 00:30:27.617 "num_base_bdevs_operational": 4, 00:30:27.617 "base_bdevs_list": [ 00:30:27.617 { 00:30:27.617 "name": "NewBaseBdev", 00:30:27.617 "uuid": "9f9bdb8c-da17-4dd7-b75a-b48572e70654", 00:30:27.617 "is_configured": true, 00:30:27.617 "data_offset": 0, 00:30:27.617 "data_size": 65536 00:30:27.617 }, 00:30:27.617 { 00:30:27.617 "name": "BaseBdev2", 00:30:27.617 "uuid": "e294b34e-eee7-4477-95fc-b59c454dc224", 00:30:27.617 "is_configured": true, 00:30:27.617 "data_offset": 0, 00:30:27.617 "data_size": 65536 00:30:27.617 }, 00:30:27.617 { 00:30:27.617 "name": "BaseBdev3", 00:30:27.617 "uuid": "e31bb439-0414-42b0-96ed-8319b3f03ad8", 00:30:27.617 "is_configured": true, 00:30:27.617 "data_offset": 0, 00:30:27.617 "data_size": 65536 00:30:27.617 }, 00:30:27.617 { 00:30:27.617 "name": "BaseBdev4", 00:30:27.617 "uuid": "8f73e27d-3d29-43e9-bf90-5b8af17f7ed2", 00:30:27.617 "is_configured": true, 00:30:27.617 "data_offset": 0, 00:30:27.617 "data_size": 65536 00:30:27.617 } 00:30:27.617 ] 00:30:27.617 } 00:30:27.617 } 00:30:27.617 }' 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:30:27.617 BaseBdev2 00:30:27.617 BaseBdev3 00:30:27.617 BaseBdev4' 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:27.617 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:27.876 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:27.876 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:27.876 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:27.876 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:27.876 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:30:27.876 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:27.876 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:27.876 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:27.876 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:27.876 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:27.876 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:27.876 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:30:27.876 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:27.876 20:29:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:27.876 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:27.876 20:29:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:27.876 20:29:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:27.876 20:29:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:27.876 20:29:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:27.876 20:29:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:30:27.876 20:29:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:27.876 20:29:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:27.876 20:29:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:27.876 20:29:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:27.876 20:29:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:27.876 20:29:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:27.876 20:29:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:30:27.876 20:29:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:27.876 20:29:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:27.876 [2024-10-01 20:29:23.073810] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:30:27.876 [2024-10-01 20:29:23.074065] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:27.876 [2024-10-01 20:29:23.074276] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:27.877 [2024-10-01 20:29:23.074859] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:27.877 [2024-10-01 20:29:23.074882] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:30:27.877 20:29:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:27.877 20:29:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 84072 00:30:27.877 20:29:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 84072 ']' 00:30:27.877 20:29:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # kill -0 84072 00:30:27.877 20:29:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@955 -- # uname 00:30:27.877 20:29:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:30:27.877 20:29:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 84072 00:30:27.877 killing process with pid 84072 00:30:27.877 20:29:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:30:27.877 20:29:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:30:27.877 20:29:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 84072' 00:30:27.877 20:29:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@969 -- # kill 84072 00:30:27.877 [2024-10-01 20:29:23.116352] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:30:27.877 20:29:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@974 -- # wait 84072 00:30:28.487 [2024-10-01 20:29:23.535342] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:30:31.020 00:30:31.020 real 0m14.347s 00:30:31.020 user 0m22.619s 00:30:31.020 sys 0m2.164s 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:30:31.020 ************************************ 00:30:31.020 END TEST raid5f_state_function_test 00:30:31.020 ************************************ 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:30:31.020 20:29:25 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 4 true 00:30:31.020 20:29:25 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:30:31.020 20:29:25 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:30:31.020 20:29:25 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:30:31.020 ************************************ 00:30:31.020 START TEST raid5f_state_function_test_sb 00:30:31.020 ************************************ 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid5f 4 true 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:30:31.020 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=84765 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:30:31.021 Process raid pid: 84765 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 84765' 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 84765 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 84765 ']' 00:30:31.021 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:30:31.021 20:29:25 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:31.021 [2024-10-01 20:29:25.896378] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:30:31.021 [2024-10-01 20:29:25.897549] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:30:31.021 [2024-10-01 20:29:26.072648] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:30:31.280 [2024-10-01 20:29:26.357734] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:30:31.540 [2024-10-01 20:29:26.612737] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:31.540 [2024-10-01 20:29:26.612848] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:32.108 [2024-10-01 20:29:27.114847] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:30:32.108 [2024-10-01 20:29:27.115155] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:30:32.108 [2024-10-01 20:29:27.115189] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:30:32.108 [2024-10-01 20:29:27.115217] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:30:32.108 [2024-10-01 20:29:27.115232] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:30:32.108 [2024-10-01 20:29:27.115252] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:30:32.108 [2024-10-01 20:29:27.115265] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:30:32.108 [2024-10-01 20:29:27.115285] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:32.108 "name": "Existed_Raid", 00:30:32.108 "uuid": "977d8b14-7b5f-456a-91be-482434f3d360", 00:30:32.108 "strip_size_kb": 64, 00:30:32.108 "state": "configuring", 00:30:32.108 "raid_level": "raid5f", 00:30:32.108 "superblock": true, 00:30:32.108 "num_base_bdevs": 4, 00:30:32.108 "num_base_bdevs_discovered": 0, 00:30:32.108 "num_base_bdevs_operational": 4, 00:30:32.108 "base_bdevs_list": [ 00:30:32.108 { 00:30:32.108 "name": "BaseBdev1", 00:30:32.108 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:32.108 "is_configured": false, 00:30:32.108 "data_offset": 0, 00:30:32.108 "data_size": 0 00:30:32.108 }, 00:30:32.108 { 00:30:32.108 "name": "BaseBdev2", 00:30:32.108 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:32.108 "is_configured": false, 00:30:32.108 "data_offset": 0, 00:30:32.108 "data_size": 0 00:30:32.108 }, 00:30:32.108 { 00:30:32.108 "name": "BaseBdev3", 00:30:32.108 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:32.108 "is_configured": false, 00:30:32.108 "data_offset": 0, 00:30:32.108 "data_size": 0 00:30:32.108 }, 00:30:32.108 { 00:30:32.108 "name": "BaseBdev4", 00:30:32.108 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:32.108 "is_configured": false, 00:30:32.108 "data_offset": 0, 00:30:32.108 "data_size": 0 00:30:32.108 } 00:30:32.108 ] 00:30:32.108 }' 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:32.108 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:32.676 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:30:32.676 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:32.676 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:32.676 [2024-10-01 20:29:27.715020] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:30:32.676 [2024-10-01 20:29:27.715126] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:30:32.676 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:32.676 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:30:32.676 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:32.676 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:32.676 [2024-10-01 20:29:27.723001] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:30:32.676 [2024-10-01 20:29:27.723067] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:30:32.676 [2024-10-01 20:29:27.723102] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:30:32.676 [2024-10-01 20:29:27.723124] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:30:32.676 [2024-10-01 20:29:27.723148] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:30:32.676 [2024-10-01 20:29:27.723167] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:30:32.676 [2024-10-01 20:29:27.723196] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:30:32.676 [2024-10-01 20:29:27.723216] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:30:32.676 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:32.676 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:30:32.676 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:32.676 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:32.676 [2024-10-01 20:29:27.777568] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:30:32.676 BaseBdev1 00:30:32.676 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:32.676 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:30:32.676 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:32.677 [ 00:30:32.677 { 00:30:32.677 "name": "BaseBdev1", 00:30:32.677 "aliases": [ 00:30:32.677 "d11fc09f-9e1c-4cd6-9b91-12c8b5bae385" 00:30:32.677 ], 00:30:32.677 "product_name": "Malloc disk", 00:30:32.677 "block_size": 512, 00:30:32.677 "num_blocks": 65536, 00:30:32.677 "uuid": "d11fc09f-9e1c-4cd6-9b91-12c8b5bae385", 00:30:32.677 "assigned_rate_limits": { 00:30:32.677 "rw_ios_per_sec": 0, 00:30:32.677 "rw_mbytes_per_sec": 0, 00:30:32.677 "r_mbytes_per_sec": 0, 00:30:32.677 "w_mbytes_per_sec": 0 00:30:32.677 }, 00:30:32.677 "claimed": true, 00:30:32.677 "claim_type": "exclusive_write", 00:30:32.677 "zoned": false, 00:30:32.677 "supported_io_types": { 00:30:32.677 "read": true, 00:30:32.677 "write": true, 00:30:32.677 "unmap": true, 00:30:32.677 "flush": true, 00:30:32.677 "reset": true, 00:30:32.677 "nvme_admin": false, 00:30:32.677 "nvme_io": false, 00:30:32.677 "nvme_io_md": false, 00:30:32.677 "write_zeroes": true, 00:30:32.677 "zcopy": true, 00:30:32.677 "get_zone_info": false, 00:30:32.677 "zone_management": false, 00:30:32.677 "zone_append": false, 00:30:32.677 "compare": false, 00:30:32.677 "compare_and_write": false, 00:30:32.677 "abort": true, 00:30:32.677 "seek_hole": false, 00:30:32.677 "seek_data": false, 00:30:32.677 "copy": true, 00:30:32.677 "nvme_iov_md": false 00:30:32.677 }, 00:30:32.677 "memory_domains": [ 00:30:32.677 { 00:30:32.677 "dma_device_id": "system", 00:30:32.677 "dma_device_type": 1 00:30:32.677 }, 00:30:32.677 { 00:30:32.677 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:32.677 "dma_device_type": 2 00:30:32.677 } 00:30:32.677 ], 00:30:32.677 "driver_specific": {} 00:30:32.677 } 00:30:32.677 ] 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:32.677 "name": "Existed_Raid", 00:30:32.677 "uuid": "163544ad-53e4-41b5-8373-33a0db7abe2e", 00:30:32.677 "strip_size_kb": 64, 00:30:32.677 "state": "configuring", 00:30:32.677 "raid_level": "raid5f", 00:30:32.677 "superblock": true, 00:30:32.677 "num_base_bdevs": 4, 00:30:32.677 "num_base_bdevs_discovered": 1, 00:30:32.677 "num_base_bdevs_operational": 4, 00:30:32.677 "base_bdevs_list": [ 00:30:32.677 { 00:30:32.677 "name": "BaseBdev1", 00:30:32.677 "uuid": "d11fc09f-9e1c-4cd6-9b91-12c8b5bae385", 00:30:32.677 "is_configured": true, 00:30:32.677 "data_offset": 2048, 00:30:32.677 "data_size": 63488 00:30:32.677 }, 00:30:32.677 { 00:30:32.677 "name": "BaseBdev2", 00:30:32.677 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:32.677 "is_configured": false, 00:30:32.677 "data_offset": 0, 00:30:32.677 "data_size": 0 00:30:32.677 }, 00:30:32.677 { 00:30:32.677 "name": "BaseBdev3", 00:30:32.677 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:32.677 "is_configured": false, 00:30:32.677 "data_offset": 0, 00:30:32.677 "data_size": 0 00:30:32.677 }, 00:30:32.677 { 00:30:32.677 "name": "BaseBdev4", 00:30:32.677 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:32.677 "is_configured": false, 00:30:32.677 "data_offset": 0, 00:30:32.677 "data_size": 0 00:30:32.677 } 00:30:32.677 ] 00:30:32.677 }' 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:32.677 20:29:27 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:33.246 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:30:33.246 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:33.246 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:33.246 [2024-10-01 20:29:28.377955] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:30:33.246 [2024-10-01 20:29:28.378607] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:30:33.246 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:33.246 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:30:33.246 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:33.246 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:33.246 [2024-10-01 20:29:28.390009] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:30:33.247 [2024-10-01 20:29:28.393125] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:30:33.247 [2024-10-01 20:29:28.393416] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:30:33.247 [2024-10-01 20:29:28.393587] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:30:33.247 [2024-10-01 20:29:28.393787] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:30:33.247 [2024-10-01 20:29:28.393966] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:30:33.247 [2024-10-01 20:29:28.394039] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:33.247 "name": "Existed_Raid", 00:30:33.247 "uuid": "5160799d-da17-4fc7-abe2-6b19e83bedd2", 00:30:33.247 "strip_size_kb": 64, 00:30:33.247 "state": "configuring", 00:30:33.247 "raid_level": "raid5f", 00:30:33.247 "superblock": true, 00:30:33.247 "num_base_bdevs": 4, 00:30:33.247 "num_base_bdevs_discovered": 1, 00:30:33.247 "num_base_bdevs_operational": 4, 00:30:33.247 "base_bdevs_list": [ 00:30:33.247 { 00:30:33.247 "name": "BaseBdev1", 00:30:33.247 "uuid": "d11fc09f-9e1c-4cd6-9b91-12c8b5bae385", 00:30:33.247 "is_configured": true, 00:30:33.247 "data_offset": 2048, 00:30:33.247 "data_size": 63488 00:30:33.247 }, 00:30:33.247 { 00:30:33.247 "name": "BaseBdev2", 00:30:33.247 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:33.247 "is_configured": false, 00:30:33.247 "data_offset": 0, 00:30:33.247 "data_size": 0 00:30:33.247 }, 00:30:33.247 { 00:30:33.247 "name": "BaseBdev3", 00:30:33.247 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:33.247 "is_configured": false, 00:30:33.247 "data_offset": 0, 00:30:33.247 "data_size": 0 00:30:33.247 }, 00:30:33.247 { 00:30:33.247 "name": "BaseBdev4", 00:30:33.247 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:33.247 "is_configured": false, 00:30:33.247 "data_offset": 0, 00:30:33.247 "data_size": 0 00:30:33.247 } 00:30:33.247 ] 00:30:33.247 }' 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:33.247 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:33.815 [2024-10-01 20:29:28.968952] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:33.815 BaseBdev2 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:33.815 20:29:28 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:33.815 [ 00:30:33.815 { 00:30:33.815 "name": "BaseBdev2", 00:30:33.815 "aliases": [ 00:30:33.815 "74306f1e-f393-4219-b050-837f9a02ce71" 00:30:33.815 ], 00:30:33.815 "product_name": "Malloc disk", 00:30:33.815 "block_size": 512, 00:30:33.815 "num_blocks": 65536, 00:30:33.815 "uuid": "74306f1e-f393-4219-b050-837f9a02ce71", 00:30:33.815 "assigned_rate_limits": { 00:30:33.815 "rw_ios_per_sec": 0, 00:30:33.815 "rw_mbytes_per_sec": 0, 00:30:33.815 "r_mbytes_per_sec": 0, 00:30:33.815 "w_mbytes_per_sec": 0 00:30:33.815 }, 00:30:33.815 "claimed": true, 00:30:33.815 "claim_type": "exclusive_write", 00:30:33.815 "zoned": false, 00:30:33.815 "supported_io_types": { 00:30:33.815 "read": true, 00:30:33.816 "write": true, 00:30:33.816 "unmap": true, 00:30:33.816 "flush": true, 00:30:33.816 "reset": true, 00:30:33.816 "nvme_admin": false, 00:30:33.816 "nvme_io": false, 00:30:33.816 "nvme_io_md": false, 00:30:33.816 "write_zeroes": true, 00:30:33.816 "zcopy": true, 00:30:33.816 "get_zone_info": false, 00:30:33.816 "zone_management": false, 00:30:33.816 "zone_append": false, 00:30:33.816 "compare": false, 00:30:33.816 "compare_and_write": false, 00:30:33.816 "abort": true, 00:30:33.816 "seek_hole": false, 00:30:33.816 "seek_data": false, 00:30:33.816 "copy": true, 00:30:33.816 "nvme_iov_md": false 00:30:33.816 }, 00:30:33.816 "memory_domains": [ 00:30:33.816 { 00:30:33.816 "dma_device_id": "system", 00:30:33.816 "dma_device_type": 1 00:30:33.816 }, 00:30:33.816 { 00:30:33.816 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:33.816 "dma_device_type": 2 00:30:33.816 } 00:30:33.816 ], 00:30:33.816 "driver_specific": {} 00:30:33.816 } 00:30:33.816 ] 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:33.816 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:34.076 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:34.076 "name": "Existed_Raid", 00:30:34.076 "uuid": "5160799d-da17-4fc7-abe2-6b19e83bedd2", 00:30:34.076 "strip_size_kb": 64, 00:30:34.076 "state": "configuring", 00:30:34.076 "raid_level": "raid5f", 00:30:34.076 "superblock": true, 00:30:34.076 "num_base_bdevs": 4, 00:30:34.076 "num_base_bdevs_discovered": 2, 00:30:34.076 "num_base_bdevs_operational": 4, 00:30:34.076 "base_bdevs_list": [ 00:30:34.076 { 00:30:34.076 "name": "BaseBdev1", 00:30:34.077 "uuid": "d11fc09f-9e1c-4cd6-9b91-12c8b5bae385", 00:30:34.077 "is_configured": true, 00:30:34.077 "data_offset": 2048, 00:30:34.077 "data_size": 63488 00:30:34.077 }, 00:30:34.077 { 00:30:34.077 "name": "BaseBdev2", 00:30:34.077 "uuid": "74306f1e-f393-4219-b050-837f9a02ce71", 00:30:34.077 "is_configured": true, 00:30:34.077 "data_offset": 2048, 00:30:34.077 "data_size": 63488 00:30:34.077 }, 00:30:34.077 { 00:30:34.077 "name": "BaseBdev3", 00:30:34.077 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:34.077 "is_configured": false, 00:30:34.077 "data_offset": 0, 00:30:34.077 "data_size": 0 00:30:34.077 }, 00:30:34.077 { 00:30:34.077 "name": "BaseBdev4", 00:30:34.077 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:34.077 "is_configured": false, 00:30:34.077 "data_offset": 0, 00:30:34.077 "data_size": 0 00:30:34.077 } 00:30:34.077 ] 00:30:34.077 }' 00:30:34.077 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:34.077 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:34.336 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:30:34.336 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:34.336 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:34.595 [2024-10-01 20:29:29.628703] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:30:34.595 BaseBdev3 00:30:34.595 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:34.595 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:34.596 [ 00:30:34.596 { 00:30:34.596 "name": "BaseBdev3", 00:30:34.596 "aliases": [ 00:30:34.596 "ce599d29-8f4b-460c-9f70-c836e4b8ee35" 00:30:34.596 ], 00:30:34.596 "product_name": "Malloc disk", 00:30:34.596 "block_size": 512, 00:30:34.596 "num_blocks": 65536, 00:30:34.596 "uuid": "ce599d29-8f4b-460c-9f70-c836e4b8ee35", 00:30:34.596 "assigned_rate_limits": { 00:30:34.596 "rw_ios_per_sec": 0, 00:30:34.596 "rw_mbytes_per_sec": 0, 00:30:34.596 "r_mbytes_per_sec": 0, 00:30:34.596 "w_mbytes_per_sec": 0 00:30:34.596 }, 00:30:34.596 "claimed": true, 00:30:34.596 "claim_type": "exclusive_write", 00:30:34.596 "zoned": false, 00:30:34.596 "supported_io_types": { 00:30:34.596 "read": true, 00:30:34.596 "write": true, 00:30:34.596 "unmap": true, 00:30:34.596 "flush": true, 00:30:34.596 "reset": true, 00:30:34.596 "nvme_admin": false, 00:30:34.596 "nvme_io": false, 00:30:34.596 "nvme_io_md": false, 00:30:34.596 "write_zeroes": true, 00:30:34.596 "zcopy": true, 00:30:34.596 "get_zone_info": false, 00:30:34.596 "zone_management": false, 00:30:34.596 "zone_append": false, 00:30:34.596 "compare": false, 00:30:34.596 "compare_and_write": false, 00:30:34.596 "abort": true, 00:30:34.596 "seek_hole": false, 00:30:34.596 "seek_data": false, 00:30:34.596 "copy": true, 00:30:34.596 "nvme_iov_md": false 00:30:34.596 }, 00:30:34.596 "memory_domains": [ 00:30:34.596 { 00:30:34.596 "dma_device_id": "system", 00:30:34.596 "dma_device_type": 1 00:30:34.596 }, 00:30:34.596 { 00:30:34.596 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:34.596 "dma_device_type": 2 00:30:34.596 } 00:30:34.596 ], 00:30:34.596 "driver_specific": {} 00:30:34.596 } 00:30:34.596 ] 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:34.596 "name": "Existed_Raid", 00:30:34.596 "uuid": "5160799d-da17-4fc7-abe2-6b19e83bedd2", 00:30:34.596 "strip_size_kb": 64, 00:30:34.596 "state": "configuring", 00:30:34.596 "raid_level": "raid5f", 00:30:34.596 "superblock": true, 00:30:34.596 "num_base_bdevs": 4, 00:30:34.596 "num_base_bdevs_discovered": 3, 00:30:34.596 "num_base_bdevs_operational": 4, 00:30:34.596 "base_bdevs_list": [ 00:30:34.596 { 00:30:34.596 "name": "BaseBdev1", 00:30:34.596 "uuid": "d11fc09f-9e1c-4cd6-9b91-12c8b5bae385", 00:30:34.596 "is_configured": true, 00:30:34.596 "data_offset": 2048, 00:30:34.596 "data_size": 63488 00:30:34.596 }, 00:30:34.596 { 00:30:34.596 "name": "BaseBdev2", 00:30:34.596 "uuid": "74306f1e-f393-4219-b050-837f9a02ce71", 00:30:34.596 "is_configured": true, 00:30:34.596 "data_offset": 2048, 00:30:34.596 "data_size": 63488 00:30:34.596 }, 00:30:34.596 { 00:30:34.596 "name": "BaseBdev3", 00:30:34.596 "uuid": "ce599d29-8f4b-460c-9f70-c836e4b8ee35", 00:30:34.596 "is_configured": true, 00:30:34.596 "data_offset": 2048, 00:30:34.596 "data_size": 63488 00:30:34.596 }, 00:30:34.596 { 00:30:34.596 "name": "BaseBdev4", 00:30:34.596 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:34.596 "is_configured": false, 00:30:34.596 "data_offset": 0, 00:30:34.596 "data_size": 0 00:30:34.596 } 00:30:34.596 ] 00:30:34.596 }' 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:34.596 20:29:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.165 [2024-10-01 20:29:30.274335] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:30:35.165 [2024-10-01 20:29:30.275030] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:30:35.165 [2024-10-01 20:29:30.275065] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:30:35.165 BaseBdev4 00:30:35.165 [2024-10-01 20:29:30.275466] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.165 [2024-10-01 20:29:30.283512] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:30:35.165 [2024-10-01 20:29:30.283546] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:30:35.165 [2024-10-01 20:29:30.284246] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.165 [ 00:30:35.165 { 00:30:35.165 "name": "BaseBdev4", 00:30:35.165 "aliases": [ 00:30:35.165 "bd206acf-ef74-4138-9ee4-91fcd7509c49" 00:30:35.165 ], 00:30:35.165 "product_name": "Malloc disk", 00:30:35.165 "block_size": 512, 00:30:35.165 "num_blocks": 65536, 00:30:35.165 "uuid": "bd206acf-ef74-4138-9ee4-91fcd7509c49", 00:30:35.165 "assigned_rate_limits": { 00:30:35.165 "rw_ios_per_sec": 0, 00:30:35.165 "rw_mbytes_per_sec": 0, 00:30:35.165 "r_mbytes_per_sec": 0, 00:30:35.165 "w_mbytes_per_sec": 0 00:30:35.165 }, 00:30:35.165 "claimed": true, 00:30:35.165 "claim_type": "exclusive_write", 00:30:35.165 "zoned": false, 00:30:35.165 "supported_io_types": { 00:30:35.165 "read": true, 00:30:35.165 "write": true, 00:30:35.165 "unmap": true, 00:30:35.165 "flush": true, 00:30:35.165 "reset": true, 00:30:35.165 "nvme_admin": false, 00:30:35.165 "nvme_io": false, 00:30:35.165 "nvme_io_md": false, 00:30:35.165 "write_zeroes": true, 00:30:35.165 "zcopy": true, 00:30:35.165 "get_zone_info": false, 00:30:35.165 "zone_management": false, 00:30:35.165 "zone_append": false, 00:30:35.165 "compare": false, 00:30:35.165 "compare_and_write": false, 00:30:35.165 "abort": true, 00:30:35.165 "seek_hole": false, 00:30:35.165 "seek_data": false, 00:30:35.165 "copy": true, 00:30:35.165 "nvme_iov_md": false 00:30:35.165 }, 00:30:35.165 "memory_domains": [ 00:30:35.165 { 00:30:35.165 "dma_device_id": "system", 00:30:35.165 "dma_device_type": 1 00:30:35.165 }, 00:30:35.165 { 00:30:35.165 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:35.165 "dma_device_type": 2 00:30:35.165 } 00:30:35.165 ], 00:30:35.165 "driver_specific": {} 00:30:35.165 } 00:30:35.165 ] 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:35.165 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:35.165 "name": "Existed_Raid", 00:30:35.165 "uuid": "5160799d-da17-4fc7-abe2-6b19e83bedd2", 00:30:35.165 "strip_size_kb": 64, 00:30:35.165 "state": "online", 00:30:35.165 "raid_level": "raid5f", 00:30:35.165 "superblock": true, 00:30:35.165 "num_base_bdevs": 4, 00:30:35.165 "num_base_bdevs_discovered": 4, 00:30:35.165 "num_base_bdevs_operational": 4, 00:30:35.165 "base_bdevs_list": [ 00:30:35.165 { 00:30:35.165 "name": "BaseBdev1", 00:30:35.165 "uuid": "d11fc09f-9e1c-4cd6-9b91-12c8b5bae385", 00:30:35.165 "is_configured": true, 00:30:35.165 "data_offset": 2048, 00:30:35.165 "data_size": 63488 00:30:35.165 }, 00:30:35.165 { 00:30:35.165 "name": "BaseBdev2", 00:30:35.165 "uuid": "74306f1e-f393-4219-b050-837f9a02ce71", 00:30:35.165 "is_configured": true, 00:30:35.165 "data_offset": 2048, 00:30:35.165 "data_size": 63488 00:30:35.165 }, 00:30:35.165 { 00:30:35.166 "name": "BaseBdev3", 00:30:35.166 "uuid": "ce599d29-8f4b-460c-9f70-c836e4b8ee35", 00:30:35.166 "is_configured": true, 00:30:35.166 "data_offset": 2048, 00:30:35.166 "data_size": 63488 00:30:35.166 }, 00:30:35.166 { 00:30:35.166 "name": "BaseBdev4", 00:30:35.166 "uuid": "bd206acf-ef74-4138-9ee4-91fcd7509c49", 00:30:35.166 "is_configured": true, 00:30:35.166 "data_offset": 2048, 00:30:35.166 "data_size": 63488 00:30:35.166 } 00:30:35.166 ] 00:30:35.166 }' 00:30:35.166 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:35.166 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.734 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:30:35.734 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:30:35.734 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:30:35.734 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:30:35.734 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:30:35.734 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:30:35.734 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:30:35.734 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:35.734 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:30:35.734 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.734 [2024-10-01 20:29:30.877306] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:35.734 20:29:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:35.734 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:30:35.734 "name": "Existed_Raid", 00:30:35.734 "aliases": [ 00:30:35.734 "5160799d-da17-4fc7-abe2-6b19e83bedd2" 00:30:35.734 ], 00:30:35.734 "product_name": "Raid Volume", 00:30:35.734 "block_size": 512, 00:30:35.734 "num_blocks": 190464, 00:30:35.734 "uuid": "5160799d-da17-4fc7-abe2-6b19e83bedd2", 00:30:35.734 "assigned_rate_limits": { 00:30:35.734 "rw_ios_per_sec": 0, 00:30:35.734 "rw_mbytes_per_sec": 0, 00:30:35.734 "r_mbytes_per_sec": 0, 00:30:35.734 "w_mbytes_per_sec": 0 00:30:35.734 }, 00:30:35.734 "claimed": false, 00:30:35.734 "zoned": false, 00:30:35.734 "supported_io_types": { 00:30:35.734 "read": true, 00:30:35.734 "write": true, 00:30:35.734 "unmap": false, 00:30:35.734 "flush": false, 00:30:35.734 "reset": true, 00:30:35.734 "nvme_admin": false, 00:30:35.734 "nvme_io": false, 00:30:35.734 "nvme_io_md": false, 00:30:35.734 "write_zeroes": true, 00:30:35.734 "zcopy": false, 00:30:35.734 "get_zone_info": false, 00:30:35.734 "zone_management": false, 00:30:35.734 "zone_append": false, 00:30:35.734 "compare": false, 00:30:35.734 "compare_and_write": false, 00:30:35.734 "abort": false, 00:30:35.734 "seek_hole": false, 00:30:35.734 "seek_data": false, 00:30:35.734 "copy": false, 00:30:35.734 "nvme_iov_md": false 00:30:35.734 }, 00:30:35.734 "driver_specific": { 00:30:35.734 "raid": { 00:30:35.734 "uuid": "5160799d-da17-4fc7-abe2-6b19e83bedd2", 00:30:35.734 "strip_size_kb": 64, 00:30:35.734 "state": "online", 00:30:35.734 "raid_level": "raid5f", 00:30:35.734 "superblock": true, 00:30:35.734 "num_base_bdevs": 4, 00:30:35.734 "num_base_bdevs_discovered": 4, 00:30:35.734 "num_base_bdevs_operational": 4, 00:30:35.734 "base_bdevs_list": [ 00:30:35.734 { 00:30:35.734 "name": "BaseBdev1", 00:30:35.734 "uuid": "d11fc09f-9e1c-4cd6-9b91-12c8b5bae385", 00:30:35.734 "is_configured": true, 00:30:35.734 "data_offset": 2048, 00:30:35.734 "data_size": 63488 00:30:35.734 }, 00:30:35.734 { 00:30:35.734 "name": "BaseBdev2", 00:30:35.734 "uuid": "74306f1e-f393-4219-b050-837f9a02ce71", 00:30:35.734 "is_configured": true, 00:30:35.734 "data_offset": 2048, 00:30:35.734 "data_size": 63488 00:30:35.734 }, 00:30:35.734 { 00:30:35.734 "name": "BaseBdev3", 00:30:35.734 "uuid": "ce599d29-8f4b-460c-9f70-c836e4b8ee35", 00:30:35.734 "is_configured": true, 00:30:35.734 "data_offset": 2048, 00:30:35.734 "data_size": 63488 00:30:35.734 }, 00:30:35.734 { 00:30:35.734 "name": "BaseBdev4", 00:30:35.734 "uuid": "bd206acf-ef74-4138-9ee4-91fcd7509c49", 00:30:35.734 "is_configured": true, 00:30:35.734 "data_offset": 2048, 00:30:35.734 "data_size": 63488 00:30:35.734 } 00:30:35.734 ] 00:30:35.734 } 00:30:35.734 } 00:30:35.734 }' 00:30:35.734 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:30:35.734 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:30:35.734 BaseBdev2 00:30:35.734 BaseBdev3 00:30:35.734 BaseBdev4' 00:30:35.734 20:29:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:35.994 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:36.283 [2024-10-01 20:29:31.253320] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:36.283 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:36.283 "name": "Existed_Raid", 00:30:36.283 "uuid": "5160799d-da17-4fc7-abe2-6b19e83bedd2", 00:30:36.283 "strip_size_kb": 64, 00:30:36.283 "state": "online", 00:30:36.283 "raid_level": "raid5f", 00:30:36.283 "superblock": true, 00:30:36.283 "num_base_bdevs": 4, 00:30:36.283 "num_base_bdevs_discovered": 3, 00:30:36.283 "num_base_bdevs_operational": 3, 00:30:36.283 "base_bdevs_list": [ 00:30:36.283 { 00:30:36.283 "name": null, 00:30:36.283 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:36.283 "is_configured": false, 00:30:36.283 "data_offset": 0, 00:30:36.283 "data_size": 63488 00:30:36.283 }, 00:30:36.283 { 00:30:36.283 "name": "BaseBdev2", 00:30:36.283 "uuid": "74306f1e-f393-4219-b050-837f9a02ce71", 00:30:36.283 "is_configured": true, 00:30:36.283 "data_offset": 2048, 00:30:36.283 "data_size": 63488 00:30:36.283 }, 00:30:36.283 { 00:30:36.283 "name": "BaseBdev3", 00:30:36.283 "uuid": "ce599d29-8f4b-460c-9f70-c836e4b8ee35", 00:30:36.283 "is_configured": true, 00:30:36.283 "data_offset": 2048, 00:30:36.284 "data_size": 63488 00:30:36.284 }, 00:30:36.284 { 00:30:36.284 "name": "BaseBdev4", 00:30:36.284 "uuid": "bd206acf-ef74-4138-9ee4-91fcd7509c49", 00:30:36.284 "is_configured": true, 00:30:36.284 "data_offset": 2048, 00:30:36.284 "data_size": 63488 00:30:36.284 } 00:30:36.284 ] 00:30:36.284 }' 00:30:36.284 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:36.284 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:36.851 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:30:36.851 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:30:36.851 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:36.851 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:36.851 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:36.851 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:30:36.851 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:36.851 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:30:36.851 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:30:36.851 20:29:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:30:36.851 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:36.851 20:29:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:36.851 [2024-10-01 20:29:31.958795] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:30:36.851 [2024-10-01 20:29:31.959040] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:36.851 [2024-10-01 20:29:32.046586] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:36.851 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:36.851 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:30:36.851 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:30:36.851 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:36.851 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:36.851 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:36.851 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:30:36.851 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:37.109 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:37.110 [2024-10-01 20:29:32.122637] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:37.110 [2024-10-01 20:29:32.268642] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:30:37.110 [2024-10-01 20:29:32.268732] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:37.110 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:37.368 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:37.368 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:30:37.368 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:30:37.368 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:30:37.368 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:30:37.368 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:30:37.368 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:30:37.368 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:37.369 BaseBdev2 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:37.369 [ 00:30:37.369 { 00:30:37.369 "name": "BaseBdev2", 00:30:37.369 "aliases": [ 00:30:37.369 "da522ba5-9489-4f6b-9b4a-6eec37492f0a" 00:30:37.369 ], 00:30:37.369 "product_name": "Malloc disk", 00:30:37.369 "block_size": 512, 00:30:37.369 "num_blocks": 65536, 00:30:37.369 "uuid": "da522ba5-9489-4f6b-9b4a-6eec37492f0a", 00:30:37.369 "assigned_rate_limits": { 00:30:37.369 "rw_ios_per_sec": 0, 00:30:37.369 "rw_mbytes_per_sec": 0, 00:30:37.369 "r_mbytes_per_sec": 0, 00:30:37.369 "w_mbytes_per_sec": 0 00:30:37.369 }, 00:30:37.369 "claimed": false, 00:30:37.369 "zoned": false, 00:30:37.369 "supported_io_types": { 00:30:37.369 "read": true, 00:30:37.369 "write": true, 00:30:37.369 "unmap": true, 00:30:37.369 "flush": true, 00:30:37.369 "reset": true, 00:30:37.369 "nvme_admin": false, 00:30:37.369 "nvme_io": false, 00:30:37.369 "nvme_io_md": false, 00:30:37.369 "write_zeroes": true, 00:30:37.369 "zcopy": true, 00:30:37.369 "get_zone_info": false, 00:30:37.369 "zone_management": false, 00:30:37.369 "zone_append": false, 00:30:37.369 "compare": false, 00:30:37.369 "compare_and_write": false, 00:30:37.369 "abort": true, 00:30:37.369 "seek_hole": false, 00:30:37.369 "seek_data": false, 00:30:37.369 "copy": true, 00:30:37.369 "nvme_iov_md": false 00:30:37.369 }, 00:30:37.369 "memory_domains": [ 00:30:37.369 { 00:30:37.369 "dma_device_id": "system", 00:30:37.369 "dma_device_type": 1 00:30:37.369 }, 00:30:37.369 { 00:30:37.369 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:37.369 "dma_device_type": 2 00:30:37.369 } 00:30:37.369 ], 00:30:37.369 "driver_specific": {} 00:30:37.369 } 00:30:37.369 ] 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:37.369 BaseBdev3 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:37.369 [ 00:30:37.369 { 00:30:37.369 "name": "BaseBdev3", 00:30:37.369 "aliases": [ 00:30:37.369 "3b30e534-78aa-411e-a8ad-3c75abeb0522" 00:30:37.369 ], 00:30:37.369 "product_name": "Malloc disk", 00:30:37.369 "block_size": 512, 00:30:37.369 "num_blocks": 65536, 00:30:37.369 "uuid": "3b30e534-78aa-411e-a8ad-3c75abeb0522", 00:30:37.369 "assigned_rate_limits": { 00:30:37.369 "rw_ios_per_sec": 0, 00:30:37.369 "rw_mbytes_per_sec": 0, 00:30:37.369 "r_mbytes_per_sec": 0, 00:30:37.369 "w_mbytes_per_sec": 0 00:30:37.369 }, 00:30:37.369 "claimed": false, 00:30:37.369 "zoned": false, 00:30:37.369 "supported_io_types": { 00:30:37.369 "read": true, 00:30:37.369 "write": true, 00:30:37.369 "unmap": true, 00:30:37.369 "flush": true, 00:30:37.369 "reset": true, 00:30:37.369 "nvme_admin": false, 00:30:37.369 "nvme_io": false, 00:30:37.369 "nvme_io_md": false, 00:30:37.369 "write_zeroes": true, 00:30:37.369 "zcopy": true, 00:30:37.369 "get_zone_info": false, 00:30:37.369 "zone_management": false, 00:30:37.369 "zone_append": false, 00:30:37.369 "compare": false, 00:30:37.369 "compare_and_write": false, 00:30:37.369 "abort": true, 00:30:37.369 "seek_hole": false, 00:30:37.369 "seek_data": false, 00:30:37.369 "copy": true, 00:30:37.369 "nvme_iov_md": false 00:30:37.369 }, 00:30:37.369 "memory_domains": [ 00:30:37.369 { 00:30:37.369 "dma_device_id": "system", 00:30:37.369 "dma_device_type": 1 00:30:37.369 }, 00:30:37.369 { 00:30:37.369 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:37.369 "dma_device_type": 2 00:30:37.369 } 00:30:37.369 ], 00:30:37.369 "driver_specific": {} 00:30:37.369 } 00:30:37.369 ] 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:37.369 BaseBdev4 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:37.369 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:37.628 [ 00:30:37.628 { 00:30:37.628 "name": "BaseBdev4", 00:30:37.628 "aliases": [ 00:30:37.628 "9234f6ae-0ad5-435c-bfc1-bab07bdf9c0e" 00:30:37.628 ], 00:30:37.628 "product_name": "Malloc disk", 00:30:37.628 "block_size": 512, 00:30:37.628 "num_blocks": 65536, 00:30:37.628 "uuid": "9234f6ae-0ad5-435c-bfc1-bab07bdf9c0e", 00:30:37.628 "assigned_rate_limits": { 00:30:37.628 "rw_ios_per_sec": 0, 00:30:37.628 "rw_mbytes_per_sec": 0, 00:30:37.628 "r_mbytes_per_sec": 0, 00:30:37.628 "w_mbytes_per_sec": 0 00:30:37.628 }, 00:30:37.628 "claimed": false, 00:30:37.628 "zoned": false, 00:30:37.628 "supported_io_types": { 00:30:37.628 "read": true, 00:30:37.628 "write": true, 00:30:37.628 "unmap": true, 00:30:37.628 "flush": true, 00:30:37.628 "reset": true, 00:30:37.628 "nvme_admin": false, 00:30:37.628 "nvme_io": false, 00:30:37.628 "nvme_io_md": false, 00:30:37.628 "write_zeroes": true, 00:30:37.628 "zcopy": true, 00:30:37.628 "get_zone_info": false, 00:30:37.628 "zone_management": false, 00:30:37.628 "zone_append": false, 00:30:37.628 "compare": false, 00:30:37.628 "compare_and_write": false, 00:30:37.628 "abort": true, 00:30:37.628 "seek_hole": false, 00:30:37.628 "seek_data": false, 00:30:37.628 "copy": true, 00:30:37.628 "nvme_iov_md": false 00:30:37.628 }, 00:30:37.628 "memory_domains": [ 00:30:37.628 { 00:30:37.628 "dma_device_id": "system", 00:30:37.628 "dma_device_type": 1 00:30:37.628 }, 00:30:37.628 { 00:30:37.628 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:37.628 "dma_device_type": 2 00:30:37.628 } 00:30:37.628 ], 00:30:37.628 "driver_specific": {} 00:30:37.628 } 00:30:37.628 ] 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:37.628 [2024-10-01 20:29:32.638491] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:30:37.628 [2024-10-01 20:29:32.638693] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:30:37.628 [2024-10-01 20:29:32.638763] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:37.628 [2024-10-01 20:29:32.641321] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:30:37.628 [2024-10-01 20:29:32.641418] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:37.628 "name": "Existed_Raid", 00:30:37.628 "uuid": "b0d7043f-d1c2-4025-a0a8-1421a0b30caf", 00:30:37.628 "strip_size_kb": 64, 00:30:37.628 "state": "configuring", 00:30:37.628 "raid_level": "raid5f", 00:30:37.628 "superblock": true, 00:30:37.628 "num_base_bdevs": 4, 00:30:37.628 "num_base_bdevs_discovered": 3, 00:30:37.628 "num_base_bdevs_operational": 4, 00:30:37.628 "base_bdevs_list": [ 00:30:37.628 { 00:30:37.628 "name": "BaseBdev1", 00:30:37.628 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:37.628 "is_configured": false, 00:30:37.628 "data_offset": 0, 00:30:37.628 "data_size": 0 00:30:37.628 }, 00:30:37.628 { 00:30:37.628 "name": "BaseBdev2", 00:30:37.628 "uuid": "da522ba5-9489-4f6b-9b4a-6eec37492f0a", 00:30:37.628 "is_configured": true, 00:30:37.628 "data_offset": 2048, 00:30:37.628 "data_size": 63488 00:30:37.628 }, 00:30:37.628 { 00:30:37.628 "name": "BaseBdev3", 00:30:37.628 "uuid": "3b30e534-78aa-411e-a8ad-3c75abeb0522", 00:30:37.628 "is_configured": true, 00:30:37.628 "data_offset": 2048, 00:30:37.628 "data_size": 63488 00:30:37.628 }, 00:30:37.628 { 00:30:37.628 "name": "BaseBdev4", 00:30:37.628 "uuid": "9234f6ae-0ad5-435c-bfc1-bab07bdf9c0e", 00:30:37.628 "is_configured": true, 00:30:37.628 "data_offset": 2048, 00:30:37.628 "data_size": 63488 00:30:37.628 } 00:30:37.628 ] 00:30:37.628 }' 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:37.628 20:29:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:38.195 [2024-10-01 20:29:33.158692] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:38.195 "name": "Existed_Raid", 00:30:38.195 "uuid": "b0d7043f-d1c2-4025-a0a8-1421a0b30caf", 00:30:38.195 "strip_size_kb": 64, 00:30:38.195 "state": "configuring", 00:30:38.195 "raid_level": "raid5f", 00:30:38.195 "superblock": true, 00:30:38.195 "num_base_bdevs": 4, 00:30:38.195 "num_base_bdevs_discovered": 2, 00:30:38.195 "num_base_bdevs_operational": 4, 00:30:38.195 "base_bdevs_list": [ 00:30:38.195 { 00:30:38.195 "name": "BaseBdev1", 00:30:38.195 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:38.195 "is_configured": false, 00:30:38.195 "data_offset": 0, 00:30:38.195 "data_size": 0 00:30:38.195 }, 00:30:38.195 { 00:30:38.195 "name": null, 00:30:38.195 "uuid": "da522ba5-9489-4f6b-9b4a-6eec37492f0a", 00:30:38.195 "is_configured": false, 00:30:38.195 "data_offset": 0, 00:30:38.195 "data_size": 63488 00:30:38.195 }, 00:30:38.195 { 00:30:38.195 "name": "BaseBdev3", 00:30:38.195 "uuid": "3b30e534-78aa-411e-a8ad-3c75abeb0522", 00:30:38.195 "is_configured": true, 00:30:38.195 "data_offset": 2048, 00:30:38.195 "data_size": 63488 00:30:38.195 }, 00:30:38.195 { 00:30:38.195 "name": "BaseBdev4", 00:30:38.195 "uuid": "9234f6ae-0ad5-435c-bfc1-bab07bdf9c0e", 00:30:38.195 "is_configured": true, 00:30:38.195 "data_offset": 2048, 00:30:38.195 "data_size": 63488 00:30:38.195 } 00:30:38.195 ] 00:30:38.195 }' 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:38.195 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:38.452 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:38.452 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:30:38.452 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:38.452 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:38.452 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:38.711 [2024-10-01 20:29:33.755132] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:30:38.711 BaseBdev1 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:38.711 [ 00:30:38.711 { 00:30:38.711 "name": "BaseBdev1", 00:30:38.711 "aliases": [ 00:30:38.711 "32dc79b9-9f91-4575-b7ed-a647fd9a37b4" 00:30:38.711 ], 00:30:38.711 "product_name": "Malloc disk", 00:30:38.711 "block_size": 512, 00:30:38.711 "num_blocks": 65536, 00:30:38.711 "uuid": "32dc79b9-9f91-4575-b7ed-a647fd9a37b4", 00:30:38.711 "assigned_rate_limits": { 00:30:38.711 "rw_ios_per_sec": 0, 00:30:38.711 "rw_mbytes_per_sec": 0, 00:30:38.711 "r_mbytes_per_sec": 0, 00:30:38.711 "w_mbytes_per_sec": 0 00:30:38.711 }, 00:30:38.711 "claimed": true, 00:30:38.711 "claim_type": "exclusive_write", 00:30:38.711 "zoned": false, 00:30:38.711 "supported_io_types": { 00:30:38.711 "read": true, 00:30:38.711 "write": true, 00:30:38.711 "unmap": true, 00:30:38.711 "flush": true, 00:30:38.711 "reset": true, 00:30:38.711 "nvme_admin": false, 00:30:38.711 "nvme_io": false, 00:30:38.711 "nvme_io_md": false, 00:30:38.711 "write_zeroes": true, 00:30:38.711 "zcopy": true, 00:30:38.711 "get_zone_info": false, 00:30:38.711 "zone_management": false, 00:30:38.711 "zone_append": false, 00:30:38.711 "compare": false, 00:30:38.711 "compare_and_write": false, 00:30:38.711 "abort": true, 00:30:38.711 "seek_hole": false, 00:30:38.711 "seek_data": false, 00:30:38.711 "copy": true, 00:30:38.711 "nvme_iov_md": false 00:30:38.711 }, 00:30:38.711 "memory_domains": [ 00:30:38.711 { 00:30:38.711 "dma_device_id": "system", 00:30:38.711 "dma_device_type": 1 00:30:38.711 }, 00:30:38.711 { 00:30:38.711 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:38.711 "dma_device_type": 2 00:30:38.711 } 00:30:38.711 ], 00:30:38.711 "driver_specific": {} 00:30:38.711 } 00:30:38.711 ] 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:38.711 "name": "Existed_Raid", 00:30:38.711 "uuid": "b0d7043f-d1c2-4025-a0a8-1421a0b30caf", 00:30:38.711 "strip_size_kb": 64, 00:30:38.711 "state": "configuring", 00:30:38.711 "raid_level": "raid5f", 00:30:38.711 "superblock": true, 00:30:38.711 "num_base_bdevs": 4, 00:30:38.711 "num_base_bdevs_discovered": 3, 00:30:38.711 "num_base_bdevs_operational": 4, 00:30:38.711 "base_bdevs_list": [ 00:30:38.711 { 00:30:38.711 "name": "BaseBdev1", 00:30:38.711 "uuid": "32dc79b9-9f91-4575-b7ed-a647fd9a37b4", 00:30:38.711 "is_configured": true, 00:30:38.711 "data_offset": 2048, 00:30:38.711 "data_size": 63488 00:30:38.711 }, 00:30:38.711 { 00:30:38.711 "name": null, 00:30:38.711 "uuid": "da522ba5-9489-4f6b-9b4a-6eec37492f0a", 00:30:38.711 "is_configured": false, 00:30:38.711 "data_offset": 0, 00:30:38.711 "data_size": 63488 00:30:38.711 }, 00:30:38.711 { 00:30:38.711 "name": "BaseBdev3", 00:30:38.711 "uuid": "3b30e534-78aa-411e-a8ad-3c75abeb0522", 00:30:38.711 "is_configured": true, 00:30:38.711 "data_offset": 2048, 00:30:38.711 "data_size": 63488 00:30:38.711 }, 00:30:38.711 { 00:30:38.711 "name": "BaseBdev4", 00:30:38.711 "uuid": "9234f6ae-0ad5-435c-bfc1-bab07bdf9c0e", 00:30:38.711 "is_configured": true, 00:30:38.711 "data_offset": 2048, 00:30:38.711 "data_size": 63488 00:30:38.711 } 00:30:38.711 ] 00:30:38.711 }' 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:38.711 20:29:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:39.279 [2024-10-01 20:29:34.367407] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:39.279 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:39.280 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:39.280 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:39.280 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:39.280 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:39.280 "name": "Existed_Raid", 00:30:39.280 "uuid": "b0d7043f-d1c2-4025-a0a8-1421a0b30caf", 00:30:39.280 "strip_size_kb": 64, 00:30:39.280 "state": "configuring", 00:30:39.280 "raid_level": "raid5f", 00:30:39.280 "superblock": true, 00:30:39.280 "num_base_bdevs": 4, 00:30:39.280 "num_base_bdevs_discovered": 2, 00:30:39.280 "num_base_bdevs_operational": 4, 00:30:39.280 "base_bdevs_list": [ 00:30:39.280 { 00:30:39.280 "name": "BaseBdev1", 00:30:39.280 "uuid": "32dc79b9-9f91-4575-b7ed-a647fd9a37b4", 00:30:39.280 "is_configured": true, 00:30:39.280 "data_offset": 2048, 00:30:39.280 "data_size": 63488 00:30:39.280 }, 00:30:39.280 { 00:30:39.280 "name": null, 00:30:39.280 "uuid": "da522ba5-9489-4f6b-9b4a-6eec37492f0a", 00:30:39.280 "is_configured": false, 00:30:39.280 "data_offset": 0, 00:30:39.280 "data_size": 63488 00:30:39.280 }, 00:30:39.280 { 00:30:39.280 "name": null, 00:30:39.280 "uuid": "3b30e534-78aa-411e-a8ad-3c75abeb0522", 00:30:39.280 "is_configured": false, 00:30:39.280 "data_offset": 0, 00:30:39.280 "data_size": 63488 00:30:39.280 }, 00:30:39.280 { 00:30:39.280 "name": "BaseBdev4", 00:30:39.280 "uuid": "9234f6ae-0ad5-435c-bfc1-bab07bdf9c0e", 00:30:39.280 "is_configured": true, 00:30:39.280 "data_offset": 2048, 00:30:39.280 "data_size": 63488 00:30:39.280 } 00:30:39.280 ] 00:30:39.280 }' 00:30:39.280 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:39.280 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:39.848 [2024-10-01 20:29:34.947649] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:39.848 20:29:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:39.848 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:39.848 "name": "Existed_Raid", 00:30:39.848 "uuid": "b0d7043f-d1c2-4025-a0a8-1421a0b30caf", 00:30:39.848 "strip_size_kb": 64, 00:30:39.848 "state": "configuring", 00:30:39.848 "raid_level": "raid5f", 00:30:39.848 "superblock": true, 00:30:39.848 "num_base_bdevs": 4, 00:30:39.848 "num_base_bdevs_discovered": 3, 00:30:39.848 "num_base_bdevs_operational": 4, 00:30:39.848 "base_bdevs_list": [ 00:30:39.848 { 00:30:39.848 "name": "BaseBdev1", 00:30:39.848 "uuid": "32dc79b9-9f91-4575-b7ed-a647fd9a37b4", 00:30:39.848 "is_configured": true, 00:30:39.848 "data_offset": 2048, 00:30:39.848 "data_size": 63488 00:30:39.848 }, 00:30:39.848 { 00:30:39.848 "name": null, 00:30:39.848 "uuid": "da522ba5-9489-4f6b-9b4a-6eec37492f0a", 00:30:39.848 "is_configured": false, 00:30:39.848 "data_offset": 0, 00:30:39.848 "data_size": 63488 00:30:39.848 }, 00:30:39.848 { 00:30:39.848 "name": "BaseBdev3", 00:30:39.848 "uuid": "3b30e534-78aa-411e-a8ad-3c75abeb0522", 00:30:39.848 "is_configured": true, 00:30:39.848 "data_offset": 2048, 00:30:39.848 "data_size": 63488 00:30:39.848 }, 00:30:39.848 { 00:30:39.848 "name": "BaseBdev4", 00:30:39.848 "uuid": "9234f6ae-0ad5-435c-bfc1-bab07bdf9c0e", 00:30:39.848 "is_configured": true, 00:30:39.848 "data_offset": 2048, 00:30:39.848 "data_size": 63488 00:30:39.848 } 00:30:39.848 ] 00:30:39.848 }' 00:30:39.848 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:39.848 20:29:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:40.415 [2024-10-01 20:29:35.503879] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:40.415 "name": "Existed_Raid", 00:30:40.415 "uuid": "b0d7043f-d1c2-4025-a0a8-1421a0b30caf", 00:30:40.415 "strip_size_kb": 64, 00:30:40.415 "state": "configuring", 00:30:40.415 "raid_level": "raid5f", 00:30:40.415 "superblock": true, 00:30:40.415 "num_base_bdevs": 4, 00:30:40.415 "num_base_bdevs_discovered": 2, 00:30:40.415 "num_base_bdevs_operational": 4, 00:30:40.415 "base_bdevs_list": [ 00:30:40.415 { 00:30:40.415 "name": null, 00:30:40.415 "uuid": "32dc79b9-9f91-4575-b7ed-a647fd9a37b4", 00:30:40.415 "is_configured": false, 00:30:40.415 "data_offset": 0, 00:30:40.415 "data_size": 63488 00:30:40.415 }, 00:30:40.415 { 00:30:40.415 "name": null, 00:30:40.415 "uuid": "da522ba5-9489-4f6b-9b4a-6eec37492f0a", 00:30:40.415 "is_configured": false, 00:30:40.415 "data_offset": 0, 00:30:40.415 "data_size": 63488 00:30:40.415 }, 00:30:40.415 { 00:30:40.415 "name": "BaseBdev3", 00:30:40.415 "uuid": "3b30e534-78aa-411e-a8ad-3c75abeb0522", 00:30:40.415 "is_configured": true, 00:30:40.415 "data_offset": 2048, 00:30:40.415 "data_size": 63488 00:30:40.415 }, 00:30:40.415 { 00:30:40.415 "name": "BaseBdev4", 00:30:40.415 "uuid": "9234f6ae-0ad5-435c-bfc1-bab07bdf9c0e", 00:30:40.415 "is_configured": true, 00:30:40.415 "data_offset": 2048, 00:30:40.415 "data_size": 63488 00:30:40.415 } 00:30:40.415 ] 00:30:40.415 }' 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:40.415 20:29:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:40.981 [2024-10-01 20:29:36.162888] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:40.981 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:40.982 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:40.982 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:40.982 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:40.982 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:40.982 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:40.982 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:40.982 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:40.982 "name": "Existed_Raid", 00:30:40.982 "uuid": "b0d7043f-d1c2-4025-a0a8-1421a0b30caf", 00:30:40.982 "strip_size_kb": 64, 00:30:40.982 "state": "configuring", 00:30:40.982 "raid_level": "raid5f", 00:30:40.982 "superblock": true, 00:30:40.982 "num_base_bdevs": 4, 00:30:40.982 "num_base_bdevs_discovered": 3, 00:30:40.982 "num_base_bdevs_operational": 4, 00:30:40.982 "base_bdevs_list": [ 00:30:40.982 { 00:30:40.982 "name": null, 00:30:40.982 "uuid": "32dc79b9-9f91-4575-b7ed-a647fd9a37b4", 00:30:40.982 "is_configured": false, 00:30:40.982 "data_offset": 0, 00:30:40.982 "data_size": 63488 00:30:40.982 }, 00:30:40.982 { 00:30:40.982 "name": "BaseBdev2", 00:30:40.982 "uuid": "da522ba5-9489-4f6b-9b4a-6eec37492f0a", 00:30:40.982 "is_configured": true, 00:30:40.982 "data_offset": 2048, 00:30:40.982 "data_size": 63488 00:30:40.982 }, 00:30:40.982 { 00:30:40.982 "name": "BaseBdev3", 00:30:40.982 "uuid": "3b30e534-78aa-411e-a8ad-3c75abeb0522", 00:30:40.982 "is_configured": true, 00:30:40.982 "data_offset": 2048, 00:30:40.982 "data_size": 63488 00:30:40.982 }, 00:30:40.982 { 00:30:40.982 "name": "BaseBdev4", 00:30:40.982 "uuid": "9234f6ae-0ad5-435c-bfc1-bab07bdf9c0e", 00:30:40.982 "is_configured": true, 00:30:40.982 "data_offset": 2048, 00:30:40.982 "data_size": 63488 00:30:40.982 } 00:30:40.982 ] 00:30:40.982 }' 00:30:40.982 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:40.982 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:41.549 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:41.549 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:41.549 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:30:41.549 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:41.549 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:41.549 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:30:41.549 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:30:41.549 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:41.549 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:41.549 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:41.549 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:41.549 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 32dc79b9-9f91-4575-b7ed-a647fd9a37b4 00:30:41.549 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:41.549 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:41.807 [2024-10-01 20:29:36.822242] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:30:41.807 [2024-10-01 20:29:36.822669] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:30:41.807 [2024-10-01 20:29:36.822691] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:30:41.807 NewBaseBdev 00:30:41.807 [2024-10-01 20:29:36.823073] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000063c0 00:30:41.807 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:41.807 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:30:41.807 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:30:41.807 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:30:41.807 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:30:41.807 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:30:41.807 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:30:41.807 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:30:41.807 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:41.807 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:41.807 [2024-10-01 20:29:36.829785] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:30:41.807 [2024-10-01 20:29:36.829825] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000008200 00:30:41.807 [2024-10-01 20:29:36.830233] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:41.807 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:41.807 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:30:41.807 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:41.807 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:41.807 [ 00:30:41.807 { 00:30:41.807 "name": "NewBaseBdev", 00:30:41.807 "aliases": [ 00:30:41.807 "32dc79b9-9f91-4575-b7ed-a647fd9a37b4" 00:30:41.807 ], 00:30:41.807 "product_name": "Malloc disk", 00:30:41.807 "block_size": 512, 00:30:41.807 "num_blocks": 65536, 00:30:41.807 "uuid": "32dc79b9-9f91-4575-b7ed-a647fd9a37b4", 00:30:41.807 "assigned_rate_limits": { 00:30:41.807 "rw_ios_per_sec": 0, 00:30:41.807 "rw_mbytes_per_sec": 0, 00:30:41.807 "r_mbytes_per_sec": 0, 00:30:41.807 "w_mbytes_per_sec": 0 00:30:41.807 }, 00:30:41.807 "claimed": true, 00:30:41.807 "claim_type": "exclusive_write", 00:30:41.807 "zoned": false, 00:30:41.807 "supported_io_types": { 00:30:41.807 "read": true, 00:30:41.807 "write": true, 00:30:41.807 "unmap": true, 00:30:41.808 "flush": true, 00:30:41.808 "reset": true, 00:30:41.808 "nvme_admin": false, 00:30:41.808 "nvme_io": false, 00:30:41.808 "nvme_io_md": false, 00:30:41.808 "write_zeroes": true, 00:30:41.808 "zcopy": true, 00:30:41.808 "get_zone_info": false, 00:30:41.808 "zone_management": false, 00:30:41.808 "zone_append": false, 00:30:41.808 "compare": false, 00:30:41.808 "compare_and_write": false, 00:30:41.808 "abort": true, 00:30:41.808 "seek_hole": false, 00:30:41.808 "seek_data": false, 00:30:41.808 "copy": true, 00:30:41.808 "nvme_iov_md": false 00:30:41.808 }, 00:30:41.808 "memory_domains": [ 00:30:41.808 { 00:30:41.808 "dma_device_id": "system", 00:30:41.808 "dma_device_type": 1 00:30:41.808 }, 00:30:41.808 { 00:30:41.808 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:30:41.808 "dma_device_type": 2 00:30:41.808 } 00:30:41.808 ], 00:30:41.808 "driver_specific": {} 00:30:41.808 } 00:30:41.808 ] 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:41.808 "name": "Existed_Raid", 00:30:41.808 "uuid": "b0d7043f-d1c2-4025-a0a8-1421a0b30caf", 00:30:41.808 "strip_size_kb": 64, 00:30:41.808 "state": "online", 00:30:41.808 "raid_level": "raid5f", 00:30:41.808 "superblock": true, 00:30:41.808 "num_base_bdevs": 4, 00:30:41.808 "num_base_bdevs_discovered": 4, 00:30:41.808 "num_base_bdevs_operational": 4, 00:30:41.808 "base_bdevs_list": [ 00:30:41.808 { 00:30:41.808 "name": "NewBaseBdev", 00:30:41.808 "uuid": "32dc79b9-9f91-4575-b7ed-a647fd9a37b4", 00:30:41.808 "is_configured": true, 00:30:41.808 "data_offset": 2048, 00:30:41.808 "data_size": 63488 00:30:41.808 }, 00:30:41.808 { 00:30:41.808 "name": "BaseBdev2", 00:30:41.808 "uuid": "da522ba5-9489-4f6b-9b4a-6eec37492f0a", 00:30:41.808 "is_configured": true, 00:30:41.808 "data_offset": 2048, 00:30:41.808 "data_size": 63488 00:30:41.808 }, 00:30:41.808 { 00:30:41.808 "name": "BaseBdev3", 00:30:41.808 "uuid": "3b30e534-78aa-411e-a8ad-3c75abeb0522", 00:30:41.808 "is_configured": true, 00:30:41.808 "data_offset": 2048, 00:30:41.808 "data_size": 63488 00:30:41.808 }, 00:30:41.808 { 00:30:41.808 "name": "BaseBdev4", 00:30:41.808 "uuid": "9234f6ae-0ad5-435c-bfc1-bab07bdf9c0e", 00:30:41.808 "is_configured": true, 00:30:41.808 "data_offset": 2048, 00:30:41.808 "data_size": 63488 00:30:41.808 } 00:30:41.808 ] 00:30:41.808 }' 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:41.808 20:29:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:42.375 [2024-10-01 20:29:37.386501] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:30:42.375 "name": "Existed_Raid", 00:30:42.375 "aliases": [ 00:30:42.375 "b0d7043f-d1c2-4025-a0a8-1421a0b30caf" 00:30:42.375 ], 00:30:42.375 "product_name": "Raid Volume", 00:30:42.375 "block_size": 512, 00:30:42.375 "num_blocks": 190464, 00:30:42.375 "uuid": "b0d7043f-d1c2-4025-a0a8-1421a0b30caf", 00:30:42.375 "assigned_rate_limits": { 00:30:42.375 "rw_ios_per_sec": 0, 00:30:42.375 "rw_mbytes_per_sec": 0, 00:30:42.375 "r_mbytes_per_sec": 0, 00:30:42.375 "w_mbytes_per_sec": 0 00:30:42.375 }, 00:30:42.375 "claimed": false, 00:30:42.375 "zoned": false, 00:30:42.375 "supported_io_types": { 00:30:42.375 "read": true, 00:30:42.375 "write": true, 00:30:42.375 "unmap": false, 00:30:42.375 "flush": false, 00:30:42.375 "reset": true, 00:30:42.375 "nvme_admin": false, 00:30:42.375 "nvme_io": false, 00:30:42.375 "nvme_io_md": false, 00:30:42.375 "write_zeroes": true, 00:30:42.375 "zcopy": false, 00:30:42.375 "get_zone_info": false, 00:30:42.375 "zone_management": false, 00:30:42.375 "zone_append": false, 00:30:42.375 "compare": false, 00:30:42.375 "compare_and_write": false, 00:30:42.375 "abort": false, 00:30:42.375 "seek_hole": false, 00:30:42.375 "seek_data": false, 00:30:42.375 "copy": false, 00:30:42.375 "nvme_iov_md": false 00:30:42.375 }, 00:30:42.375 "driver_specific": { 00:30:42.375 "raid": { 00:30:42.375 "uuid": "b0d7043f-d1c2-4025-a0a8-1421a0b30caf", 00:30:42.375 "strip_size_kb": 64, 00:30:42.375 "state": "online", 00:30:42.375 "raid_level": "raid5f", 00:30:42.375 "superblock": true, 00:30:42.375 "num_base_bdevs": 4, 00:30:42.375 "num_base_bdevs_discovered": 4, 00:30:42.375 "num_base_bdevs_operational": 4, 00:30:42.375 "base_bdevs_list": [ 00:30:42.375 { 00:30:42.375 "name": "NewBaseBdev", 00:30:42.375 "uuid": "32dc79b9-9f91-4575-b7ed-a647fd9a37b4", 00:30:42.375 "is_configured": true, 00:30:42.375 "data_offset": 2048, 00:30:42.375 "data_size": 63488 00:30:42.375 }, 00:30:42.375 { 00:30:42.375 "name": "BaseBdev2", 00:30:42.375 "uuid": "da522ba5-9489-4f6b-9b4a-6eec37492f0a", 00:30:42.375 "is_configured": true, 00:30:42.375 "data_offset": 2048, 00:30:42.375 "data_size": 63488 00:30:42.375 }, 00:30:42.375 { 00:30:42.375 "name": "BaseBdev3", 00:30:42.375 "uuid": "3b30e534-78aa-411e-a8ad-3c75abeb0522", 00:30:42.375 "is_configured": true, 00:30:42.375 "data_offset": 2048, 00:30:42.375 "data_size": 63488 00:30:42.375 }, 00:30:42.375 { 00:30:42.375 "name": "BaseBdev4", 00:30:42.375 "uuid": "9234f6ae-0ad5-435c-bfc1-bab07bdf9c0e", 00:30:42.375 "is_configured": true, 00:30:42.375 "data_offset": 2048, 00:30:42.375 "data_size": 63488 00:30:42.375 } 00:30:42.375 ] 00:30:42.375 } 00:30:42.375 } 00:30:42.375 }' 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:30:42.375 BaseBdev2 00:30:42.375 BaseBdev3 00:30:42.375 BaseBdev4' 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:42.375 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:42.633 [2024-10-01 20:29:37.762321] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:30:42.633 [2024-10-01 20:29:37.762809] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:42.633 [2024-10-01 20:29:37.762940] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:42.633 [2024-10-01 20:29:37.763447] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:42.633 [2024-10-01 20:29:37.763484] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name Existed_Raid, state offline 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 84765 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 84765 ']' 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 84765 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:30:42.633 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:30:42.634 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 84765 00:30:42.634 killing process with pid 84765 00:30:42.634 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:30:42.634 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:30:42.634 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 84765' 00:30:42.634 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 84765 00:30:42.634 [2024-10-01 20:29:37.802108] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:30:42.634 20:29:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 84765 00:30:43.200 [2024-10-01 20:29:38.213211] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:30:45.125 20:29:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:30:45.125 00:30:45.125 real 0m14.574s 00:30:45.125 user 0m22.959s 00:30:45.125 sys 0m2.223s 00:30:45.125 ************************************ 00:30:45.125 END TEST raid5f_state_function_test_sb 00:30:45.125 ************************************ 00:30:45.125 20:29:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:30:45.125 20:29:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:30:45.383 20:29:40 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 4 00:30:45.383 20:29:40 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:30:45.383 20:29:40 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:30:45.383 20:29:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:30:45.383 ************************************ 00:30:45.383 START TEST raid5f_superblock_test 00:30:45.383 ************************************ 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid5f 4 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=85460 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 85460 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 85460 ']' 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:45.383 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:30:45.383 20:29:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:45.383 [2024-10-01 20:29:40.539183] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:30:45.383 [2024-10-01 20:29:40.539631] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85460 ] 00:30:45.642 [2024-10-01 20:29:40.718658] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:30:45.900 [2024-10-01 20:29:41.008548] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:30:46.159 [2024-10-01 20:29:41.253880] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:46.159 [2024-10-01 20:29:41.254221] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:46.727 malloc1 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:46.727 [2024-10-01 20:29:41.787098] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:30:46.727 [2024-10-01 20:29:41.787424] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:46.727 [2024-10-01 20:29:41.787604] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:30:46.727 [2024-10-01 20:29:41.787794] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:46.727 [2024-10-01 20:29:41.791077] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:46.727 [2024-10-01 20:29:41.791266] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:30:46.727 pt1 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:46.727 malloc2 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:46.727 [2024-10-01 20:29:41.851050] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:30:46.727 [2024-10-01 20:29:41.851273] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:46.727 [2024-10-01 20:29:41.851370] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:30:46.727 [2024-10-01 20:29:41.851580] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:46.727 [2024-10-01 20:29:41.854597] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:46.727 [2024-10-01 20:29:41.854786] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:30:46.727 pt2 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:46.727 malloc3 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:46.727 [2024-10-01 20:29:41.913279] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:30:46.727 [2024-10-01 20:29:41.913400] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:46.727 [2024-10-01 20:29:41.913448] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:30:46.727 [2024-10-01 20:29:41.913468] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:46.727 [2024-10-01 20:29:41.916666] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:46.727 [2024-10-01 20:29:41.916747] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:30:46.727 pt3 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:46.727 malloc4 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:46.727 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:46.987 [2024-10-01 20:29:41.982555] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:30:46.987 [2024-10-01 20:29:41.982733] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:46.987 [2024-10-01 20:29:41.982785] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:30:46.987 [2024-10-01 20:29:41.982807] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:46.987 [2024-10-01 20:29:41.985948] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:46.987 [2024-10-01 20:29:41.986003] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:30:46.987 pt4 00:30:46.987 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:46.987 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:30:46.987 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:30:46.987 20:29:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:30:46.987 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:46.987 20:29:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:46.987 [2024-10-01 20:29:41.995147] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:30:46.987 [2024-10-01 20:29:41.998205] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:30:46.987 [2024-10-01 20:29:41.998408] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:30:46.987 [2024-10-01 20:29:41.998519] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:30:46.987 [2024-10-01 20:29:41.998970] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:30:46.987 [2024-10-01 20:29:41.999032] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:30:46.987 [2024-10-01 20:29:41.999577] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:30:46.987 [2024-10-01 20:29:42.008263] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:30:46.987 [2024-10-01 20:29:42.008465] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:30:46.987 [2024-10-01 20:29:42.009013] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:46.987 "name": "raid_bdev1", 00:30:46.987 "uuid": "88b7386e-e410-4597-bbaa-4f19716a85a5", 00:30:46.987 "strip_size_kb": 64, 00:30:46.987 "state": "online", 00:30:46.987 "raid_level": "raid5f", 00:30:46.987 "superblock": true, 00:30:46.987 "num_base_bdevs": 4, 00:30:46.987 "num_base_bdevs_discovered": 4, 00:30:46.987 "num_base_bdevs_operational": 4, 00:30:46.987 "base_bdevs_list": [ 00:30:46.987 { 00:30:46.987 "name": "pt1", 00:30:46.987 "uuid": "00000000-0000-0000-0000-000000000001", 00:30:46.987 "is_configured": true, 00:30:46.987 "data_offset": 2048, 00:30:46.987 "data_size": 63488 00:30:46.987 }, 00:30:46.987 { 00:30:46.987 "name": "pt2", 00:30:46.987 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:46.987 "is_configured": true, 00:30:46.987 "data_offset": 2048, 00:30:46.987 "data_size": 63488 00:30:46.987 }, 00:30:46.987 { 00:30:46.987 "name": "pt3", 00:30:46.987 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:46.987 "is_configured": true, 00:30:46.987 "data_offset": 2048, 00:30:46.987 "data_size": 63488 00:30:46.987 }, 00:30:46.987 { 00:30:46.987 "name": "pt4", 00:30:46.987 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:46.987 "is_configured": true, 00:30:46.987 "data_offset": 2048, 00:30:46.987 "data_size": 63488 00:30:46.987 } 00:30:46.987 ] 00:30:46.987 }' 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:46.987 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:47.555 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:30:47.555 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:30:47.555 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:30:47.555 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:30:47.555 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:30:47.555 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:30:47.555 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:30:47.555 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:47.555 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:30:47.555 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:47.555 [2024-10-01 20:29:42.553588] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:47.555 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:47.555 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:30:47.555 "name": "raid_bdev1", 00:30:47.555 "aliases": [ 00:30:47.555 "88b7386e-e410-4597-bbaa-4f19716a85a5" 00:30:47.555 ], 00:30:47.555 "product_name": "Raid Volume", 00:30:47.555 "block_size": 512, 00:30:47.555 "num_blocks": 190464, 00:30:47.555 "uuid": "88b7386e-e410-4597-bbaa-4f19716a85a5", 00:30:47.555 "assigned_rate_limits": { 00:30:47.555 "rw_ios_per_sec": 0, 00:30:47.555 "rw_mbytes_per_sec": 0, 00:30:47.555 "r_mbytes_per_sec": 0, 00:30:47.555 "w_mbytes_per_sec": 0 00:30:47.555 }, 00:30:47.555 "claimed": false, 00:30:47.555 "zoned": false, 00:30:47.555 "supported_io_types": { 00:30:47.555 "read": true, 00:30:47.555 "write": true, 00:30:47.555 "unmap": false, 00:30:47.555 "flush": false, 00:30:47.555 "reset": true, 00:30:47.555 "nvme_admin": false, 00:30:47.555 "nvme_io": false, 00:30:47.555 "nvme_io_md": false, 00:30:47.555 "write_zeroes": true, 00:30:47.555 "zcopy": false, 00:30:47.555 "get_zone_info": false, 00:30:47.555 "zone_management": false, 00:30:47.555 "zone_append": false, 00:30:47.556 "compare": false, 00:30:47.556 "compare_and_write": false, 00:30:47.556 "abort": false, 00:30:47.556 "seek_hole": false, 00:30:47.556 "seek_data": false, 00:30:47.556 "copy": false, 00:30:47.556 "nvme_iov_md": false 00:30:47.556 }, 00:30:47.556 "driver_specific": { 00:30:47.556 "raid": { 00:30:47.556 "uuid": "88b7386e-e410-4597-bbaa-4f19716a85a5", 00:30:47.556 "strip_size_kb": 64, 00:30:47.556 "state": "online", 00:30:47.556 "raid_level": "raid5f", 00:30:47.556 "superblock": true, 00:30:47.556 "num_base_bdevs": 4, 00:30:47.556 "num_base_bdevs_discovered": 4, 00:30:47.556 "num_base_bdevs_operational": 4, 00:30:47.556 "base_bdevs_list": [ 00:30:47.556 { 00:30:47.556 "name": "pt1", 00:30:47.556 "uuid": "00000000-0000-0000-0000-000000000001", 00:30:47.556 "is_configured": true, 00:30:47.556 "data_offset": 2048, 00:30:47.556 "data_size": 63488 00:30:47.556 }, 00:30:47.556 { 00:30:47.556 "name": "pt2", 00:30:47.556 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:47.556 "is_configured": true, 00:30:47.556 "data_offset": 2048, 00:30:47.556 "data_size": 63488 00:30:47.556 }, 00:30:47.556 { 00:30:47.556 "name": "pt3", 00:30:47.556 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:47.556 "is_configured": true, 00:30:47.556 "data_offset": 2048, 00:30:47.556 "data_size": 63488 00:30:47.556 }, 00:30:47.556 { 00:30:47.556 "name": "pt4", 00:30:47.556 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:47.556 "is_configured": true, 00:30:47.556 "data_offset": 2048, 00:30:47.556 "data_size": 63488 00:30:47.556 } 00:30:47.556 ] 00:30:47.556 } 00:30:47.556 } 00:30:47.556 }' 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:30:47.556 pt2 00:30:47.556 pt3 00:30:47.556 pt4' 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:47.556 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:30:47.814 [2024-10-01 20:29:42.941572] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=88b7386e-e410-4597-bbaa-4f19716a85a5 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 88b7386e-e410-4597-bbaa-4f19716a85a5 ']' 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:47.814 20:29:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:47.814 [2024-10-01 20:29:43.001461] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:30:47.814 [2024-10-01 20:29:43.001493] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:47.814 [2024-10-01 20:29:43.001599] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:47.814 [2024-10-01 20:29:43.001764] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:47.814 [2024-10-01 20:29:43.001821] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:30:47.814 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:47.814 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:47.814 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:47.814 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:47.814 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:30:47.814 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:47.814 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:30:47.814 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:30:47.814 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:30:47.814 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:30:47.814 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:47.814 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:30:48.074 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:48.075 [2024-10-01 20:29:43.161541] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:30:48.075 [2024-10-01 20:29:43.164267] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:30:48.075 [2024-10-01 20:29:43.164348] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:30:48.075 [2024-10-01 20:29:43.164426] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:30:48.075 [2024-10-01 20:29:43.164516] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:30:48.075 [2024-10-01 20:29:43.164606] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:30:48.075 [2024-10-01 20:29:43.164649] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:30:48.075 [2024-10-01 20:29:43.164690] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:30:48.075 [2024-10-01 20:29:43.164742] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:30:48.075 [2024-10-01 20:29:43.164773] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:30:48.075 request: 00:30:48.075 { 00:30:48.075 "name": "raid_bdev1", 00:30:48.075 "raid_level": "raid5f", 00:30:48.075 "base_bdevs": [ 00:30:48.075 "malloc1", 00:30:48.075 "malloc2", 00:30:48.075 "malloc3", 00:30:48.075 "malloc4" 00:30:48.075 ], 00:30:48.075 "strip_size_kb": 64, 00:30:48.075 "superblock": false, 00:30:48.075 "method": "bdev_raid_create", 00:30:48.075 "req_id": 1 00:30:48.075 } 00:30:48.075 Got JSON-RPC error response 00:30:48.075 response: 00:30:48.075 { 00:30:48.075 "code": -17, 00:30:48.075 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:30:48.075 } 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:48.075 [2024-10-01 20:29:43.245621] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:30:48.075 [2024-10-01 20:29:43.245899] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:48.075 [2024-10-01 20:29:43.245985] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:30:48.075 [2024-10-01 20:29:43.246153] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:48.075 [2024-10-01 20:29:43.249500] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:48.075 [2024-10-01 20:29:43.249690] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:30:48.075 [2024-10-01 20:29:43.250015] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:30:48.075 [2024-10-01 20:29:43.250240] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:30:48.075 pt1 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:48.075 "name": "raid_bdev1", 00:30:48.075 "uuid": "88b7386e-e410-4597-bbaa-4f19716a85a5", 00:30:48.075 "strip_size_kb": 64, 00:30:48.075 "state": "configuring", 00:30:48.075 "raid_level": "raid5f", 00:30:48.075 "superblock": true, 00:30:48.075 "num_base_bdevs": 4, 00:30:48.075 "num_base_bdevs_discovered": 1, 00:30:48.075 "num_base_bdevs_operational": 4, 00:30:48.075 "base_bdevs_list": [ 00:30:48.075 { 00:30:48.075 "name": "pt1", 00:30:48.075 "uuid": "00000000-0000-0000-0000-000000000001", 00:30:48.075 "is_configured": true, 00:30:48.075 "data_offset": 2048, 00:30:48.075 "data_size": 63488 00:30:48.075 }, 00:30:48.075 { 00:30:48.075 "name": null, 00:30:48.075 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:48.075 "is_configured": false, 00:30:48.075 "data_offset": 2048, 00:30:48.075 "data_size": 63488 00:30:48.075 }, 00:30:48.075 { 00:30:48.075 "name": null, 00:30:48.075 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:48.075 "is_configured": false, 00:30:48.075 "data_offset": 2048, 00:30:48.075 "data_size": 63488 00:30:48.075 }, 00:30:48.075 { 00:30:48.075 "name": null, 00:30:48.075 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:48.075 "is_configured": false, 00:30:48.075 "data_offset": 2048, 00:30:48.075 "data_size": 63488 00:30:48.075 } 00:30:48.075 ] 00:30:48.075 }' 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:48.075 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:48.642 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:30:48.642 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:30:48.642 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:48.642 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:48.642 [2024-10-01 20:29:43.782550] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:30:48.642 [2024-10-01 20:29:43.782826] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:48.642 [2024-10-01 20:29:43.782874] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:30:48.642 [2024-10-01 20:29:43.782900] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:48.642 [2024-10-01 20:29:43.783606] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:48.642 [2024-10-01 20:29:43.783655] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:30:48.642 [2024-10-01 20:29:43.783814] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:30:48.642 [2024-10-01 20:29:43.783862] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:30:48.642 pt2 00:30:48.642 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:48.642 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:48.643 [2024-10-01 20:29:43.790585] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:48.643 "name": "raid_bdev1", 00:30:48.643 "uuid": "88b7386e-e410-4597-bbaa-4f19716a85a5", 00:30:48.643 "strip_size_kb": 64, 00:30:48.643 "state": "configuring", 00:30:48.643 "raid_level": "raid5f", 00:30:48.643 "superblock": true, 00:30:48.643 "num_base_bdevs": 4, 00:30:48.643 "num_base_bdevs_discovered": 1, 00:30:48.643 "num_base_bdevs_operational": 4, 00:30:48.643 "base_bdevs_list": [ 00:30:48.643 { 00:30:48.643 "name": "pt1", 00:30:48.643 "uuid": "00000000-0000-0000-0000-000000000001", 00:30:48.643 "is_configured": true, 00:30:48.643 "data_offset": 2048, 00:30:48.643 "data_size": 63488 00:30:48.643 }, 00:30:48.643 { 00:30:48.643 "name": null, 00:30:48.643 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:48.643 "is_configured": false, 00:30:48.643 "data_offset": 0, 00:30:48.643 "data_size": 63488 00:30:48.643 }, 00:30:48.643 { 00:30:48.643 "name": null, 00:30:48.643 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:48.643 "is_configured": false, 00:30:48.643 "data_offset": 2048, 00:30:48.643 "data_size": 63488 00:30:48.643 }, 00:30:48.643 { 00:30:48.643 "name": null, 00:30:48.643 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:48.643 "is_configured": false, 00:30:48.643 "data_offset": 2048, 00:30:48.643 "data_size": 63488 00:30:48.643 } 00:30:48.643 ] 00:30:48.643 }' 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:48.643 20:29:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:49.210 [2024-10-01 20:29:44.346701] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:30:49.210 [2024-10-01 20:29:44.346967] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:49.210 [2024-10-01 20:29:44.347022] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:30:49.210 [2024-10-01 20:29:44.347043] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:49.210 [2024-10-01 20:29:44.347656] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:49.210 [2024-10-01 20:29:44.347685] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:30:49.210 [2024-10-01 20:29:44.347833] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:30:49.210 [2024-10-01 20:29:44.347872] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:30:49.210 pt2 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:49.210 [2024-10-01 20:29:44.354620] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:30:49.210 [2024-10-01 20:29:44.354842] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:49.210 [2024-10-01 20:29:44.354898] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:30:49.210 [2024-10-01 20:29:44.354920] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:49.210 [2024-10-01 20:29:44.355386] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:49.210 [2024-10-01 20:29:44.355428] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:30:49.210 [2024-10-01 20:29:44.355522] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:30:49.210 [2024-10-01 20:29:44.355566] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:30:49.210 pt3 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:49.210 [2024-10-01 20:29:44.362589] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:30:49.210 [2024-10-01 20:29:44.362799] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:49.210 [2024-10-01 20:29:44.362972] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:30:49.210 [2024-10-01 20:29:44.363105] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:49.210 [2024-10-01 20:29:44.363740] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:49.210 [2024-10-01 20:29:44.363896] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:30:49.210 [2024-10-01 20:29:44.364101] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:30:49.210 [2024-10-01 20:29:44.364248] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:30:49.210 [2024-10-01 20:29:44.364552] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:30:49.210 [2024-10-01 20:29:44.364683] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:30:49.210 [2024-10-01 20:29:44.365151] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:30:49.210 [2024-10-01 20:29:44.372042] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:30:49.210 [2024-10-01 20:29:44.372195] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:30:49.210 [2024-10-01 20:29:44.372567] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:49.210 pt4 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:49.210 "name": "raid_bdev1", 00:30:49.210 "uuid": "88b7386e-e410-4597-bbaa-4f19716a85a5", 00:30:49.210 "strip_size_kb": 64, 00:30:49.210 "state": "online", 00:30:49.210 "raid_level": "raid5f", 00:30:49.210 "superblock": true, 00:30:49.210 "num_base_bdevs": 4, 00:30:49.210 "num_base_bdevs_discovered": 4, 00:30:49.210 "num_base_bdevs_operational": 4, 00:30:49.210 "base_bdevs_list": [ 00:30:49.210 { 00:30:49.210 "name": "pt1", 00:30:49.210 "uuid": "00000000-0000-0000-0000-000000000001", 00:30:49.210 "is_configured": true, 00:30:49.210 "data_offset": 2048, 00:30:49.210 "data_size": 63488 00:30:49.210 }, 00:30:49.210 { 00:30:49.210 "name": "pt2", 00:30:49.210 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:49.210 "is_configured": true, 00:30:49.210 "data_offset": 2048, 00:30:49.210 "data_size": 63488 00:30:49.210 }, 00:30:49.210 { 00:30:49.210 "name": "pt3", 00:30:49.210 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:49.210 "is_configured": true, 00:30:49.210 "data_offset": 2048, 00:30:49.210 "data_size": 63488 00:30:49.210 }, 00:30:49.210 { 00:30:49.210 "name": "pt4", 00:30:49.210 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:49.210 "is_configured": true, 00:30:49.210 "data_offset": 2048, 00:30:49.210 "data_size": 63488 00:30:49.210 } 00:30:49.210 ] 00:30:49.210 }' 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:49.210 20:29:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:49.778 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:30:49.778 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:30:49.778 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:30:49.778 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:30:49.778 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:30:49.778 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:30:49.778 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:30:49.778 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:30:49.778 20:29:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:49.778 20:29:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:49.778 [2024-10-01 20:29:44.928580] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:49.778 20:29:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:49.778 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:30:49.778 "name": "raid_bdev1", 00:30:49.778 "aliases": [ 00:30:49.778 "88b7386e-e410-4597-bbaa-4f19716a85a5" 00:30:49.778 ], 00:30:49.778 "product_name": "Raid Volume", 00:30:49.778 "block_size": 512, 00:30:49.778 "num_blocks": 190464, 00:30:49.778 "uuid": "88b7386e-e410-4597-bbaa-4f19716a85a5", 00:30:49.778 "assigned_rate_limits": { 00:30:49.778 "rw_ios_per_sec": 0, 00:30:49.778 "rw_mbytes_per_sec": 0, 00:30:49.778 "r_mbytes_per_sec": 0, 00:30:49.778 "w_mbytes_per_sec": 0 00:30:49.778 }, 00:30:49.778 "claimed": false, 00:30:49.778 "zoned": false, 00:30:49.778 "supported_io_types": { 00:30:49.778 "read": true, 00:30:49.778 "write": true, 00:30:49.778 "unmap": false, 00:30:49.778 "flush": false, 00:30:49.778 "reset": true, 00:30:49.778 "nvme_admin": false, 00:30:49.778 "nvme_io": false, 00:30:49.778 "nvme_io_md": false, 00:30:49.778 "write_zeroes": true, 00:30:49.778 "zcopy": false, 00:30:49.778 "get_zone_info": false, 00:30:49.778 "zone_management": false, 00:30:49.778 "zone_append": false, 00:30:49.778 "compare": false, 00:30:49.778 "compare_and_write": false, 00:30:49.778 "abort": false, 00:30:49.778 "seek_hole": false, 00:30:49.778 "seek_data": false, 00:30:49.778 "copy": false, 00:30:49.778 "nvme_iov_md": false 00:30:49.778 }, 00:30:49.778 "driver_specific": { 00:30:49.778 "raid": { 00:30:49.778 "uuid": "88b7386e-e410-4597-bbaa-4f19716a85a5", 00:30:49.778 "strip_size_kb": 64, 00:30:49.778 "state": "online", 00:30:49.778 "raid_level": "raid5f", 00:30:49.778 "superblock": true, 00:30:49.778 "num_base_bdevs": 4, 00:30:49.778 "num_base_bdevs_discovered": 4, 00:30:49.778 "num_base_bdevs_operational": 4, 00:30:49.778 "base_bdevs_list": [ 00:30:49.778 { 00:30:49.778 "name": "pt1", 00:30:49.778 "uuid": "00000000-0000-0000-0000-000000000001", 00:30:49.778 "is_configured": true, 00:30:49.778 "data_offset": 2048, 00:30:49.778 "data_size": 63488 00:30:49.778 }, 00:30:49.778 { 00:30:49.778 "name": "pt2", 00:30:49.778 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:49.778 "is_configured": true, 00:30:49.778 "data_offset": 2048, 00:30:49.778 "data_size": 63488 00:30:49.778 }, 00:30:49.778 { 00:30:49.778 "name": "pt3", 00:30:49.778 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:49.778 "is_configured": true, 00:30:49.778 "data_offset": 2048, 00:30:49.778 "data_size": 63488 00:30:49.778 }, 00:30:49.778 { 00:30:49.778 "name": "pt4", 00:30:49.778 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:49.778 "is_configured": true, 00:30:49.778 "data_offset": 2048, 00:30:49.778 "data_size": 63488 00:30:49.778 } 00:30:49.778 ] 00:30:49.778 } 00:30:49.778 } 00:30:49.778 }' 00:30:49.778 20:29:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:30:49.778 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:30:49.778 pt2 00:30:49.778 pt3 00:30:49.778 pt4' 00:30:49.778 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:30:50.037 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:50.296 [2024-10-01 20:29:45.312628] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 88b7386e-e410-4597-bbaa-4f19716a85a5 '!=' 88b7386e-e410-4597-bbaa-4f19716a85a5 ']' 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:50.296 [2024-10-01 20:29:45.364459] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:50.296 "name": "raid_bdev1", 00:30:50.296 "uuid": "88b7386e-e410-4597-bbaa-4f19716a85a5", 00:30:50.296 "strip_size_kb": 64, 00:30:50.296 "state": "online", 00:30:50.296 "raid_level": "raid5f", 00:30:50.296 "superblock": true, 00:30:50.296 "num_base_bdevs": 4, 00:30:50.296 "num_base_bdevs_discovered": 3, 00:30:50.296 "num_base_bdevs_operational": 3, 00:30:50.296 "base_bdevs_list": [ 00:30:50.296 { 00:30:50.296 "name": null, 00:30:50.296 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:50.296 "is_configured": false, 00:30:50.296 "data_offset": 0, 00:30:50.296 "data_size": 63488 00:30:50.296 }, 00:30:50.296 { 00:30:50.296 "name": "pt2", 00:30:50.296 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:50.296 "is_configured": true, 00:30:50.296 "data_offset": 2048, 00:30:50.296 "data_size": 63488 00:30:50.296 }, 00:30:50.296 { 00:30:50.296 "name": "pt3", 00:30:50.296 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:50.296 "is_configured": true, 00:30:50.296 "data_offset": 2048, 00:30:50.296 "data_size": 63488 00:30:50.296 }, 00:30:50.296 { 00:30:50.296 "name": "pt4", 00:30:50.296 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:50.296 "is_configured": true, 00:30:50.296 "data_offset": 2048, 00:30:50.296 "data_size": 63488 00:30:50.296 } 00:30:50.296 ] 00:30:50.296 }' 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:50.296 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:50.864 [2024-10-01 20:29:45.916586] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:30:50.864 [2024-10-01 20:29:45.916633] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:50.864 [2024-10-01 20:29:45.916745] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:50.864 [2024-10-01 20:29:45.916888] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:50.864 [2024-10-01 20:29:45.916914] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:30:50.864 20:29:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:50.864 [2024-10-01 20:29:46.008558] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:30:50.864 [2024-10-01 20:29:46.008633] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:50.864 [2024-10-01 20:29:46.008671] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:30:50.864 [2024-10-01 20:29:46.008690] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:50.864 [2024-10-01 20:29:46.011770] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:50.864 [2024-10-01 20:29:46.011825] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:30:50.864 [2024-10-01 20:29:46.011944] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:30:50.864 [2024-10-01 20:29:46.012011] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:30:50.864 pt2 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:50.864 "name": "raid_bdev1", 00:30:50.864 "uuid": "88b7386e-e410-4597-bbaa-4f19716a85a5", 00:30:50.864 "strip_size_kb": 64, 00:30:50.864 "state": "configuring", 00:30:50.864 "raid_level": "raid5f", 00:30:50.864 "superblock": true, 00:30:50.864 "num_base_bdevs": 4, 00:30:50.864 "num_base_bdevs_discovered": 1, 00:30:50.864 "num_base_bdevs_operational": 3, 00:30:50.864 "base_bdevs_list": [ 00:30:50.864 { 00:30:50.864 "name": null, 00:30:50.864 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:50.864 "is_configured": false, 00:30:50.864 "data_offset": 2048, 00:30:50.864 "data_size": 63488 00:30:50.864 }, 00:30:50.864 { 00:30:50.864 "name": "pt2", 00:30:50.864 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:50.864 "is_configured": true, 00:30:50.864 "data_offset": 2048, 00:30:50.864 "data_size": 63488 00:30:50.864 }, 00:30:50.864 { 00:30:50.864 "name": null, 00:30:50.864 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:50.864 "is_configured": false, 00:30:50.864 "data_offset": 2048, 00:30:50.864 "data_size": 63488 00:30:50.864 }, 00:30:50.864 { 00:30:50.864 "name": null, 00:30:50.864 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:50.864 "is_configured": false, 00:30:50.864 "data_offset": 2048, 00:30:50.864 "data_size": 63488 00:30:50.864 } 00:30:50.864 ] 00:30:50.864 }' 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:50.864 20:29:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:51.430 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:30:51.430 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:30:51.430 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:30:51.430 20:29:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:51.430 20:29:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:51.430 [2024-10-01 20:29:46.524802] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:30:51.430 [2024-10-01 20:29:46.524896] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:51.430 [2024-10-01 20:29:46.524940] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:30:51.430 [2024-10-01 20:29:46.524961] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:51.430 [2024-10-01 20:29:46.525638] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:51.430 [2024-10-01 20:29:46.525686] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:30:51.430 [2024-10-01 20:29:46.525836] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:30:51.430 [2024-10-01 20:29:46.525893] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:30:51.430 pt3 00:30:51.430 20:29:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:51.430 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:30:51.430 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:51.430 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:51.430 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:51.431 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:51.431 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:51.431 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:51.431 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:51.431 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:51.431 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:51.431 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:51.431 20:29:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:51.431 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:51.431 20:29:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:51.431 20:29:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:51.431 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:51.431 "name": "raid_bdev1", 00:30:51.431 "uuid": "88b7386e-e410-4597-bbaa-4f19716a85a5", 00:30:51.431 "strip_size_kb": 64, 00:30:51.431 "state": "configuring", 00:30:51.431 "raid_level": "raid5f", 00:30:51.431 "superblock": true, 00:30:51.431 "num_base_bdevs": 4, 00:30:51.431 "num_base_bdevs_discovered": 2, 00:30:51.431 "num_base_bdevs_operational": 3, 00:30:51.431 "base_bdevs_list": [ 00:30:51.431 { 00:30:51.431 "name": null, 00:30:51.431 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:51.431 "is_configured": false, 00:30:51.431 "data_offset": 2048, 00:30:51.431 "data_size": 63488 00:30:51.431 }, 00:30:51.431 { 00:30:51.431 "name": "pt2", 00:30:51.431 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:51.431 "is_configured": true, 00:30:51.431 "data_offset": 2048, 00:30:51.431 "data_size": 63488 00:30:51.431 }, 00:30:51.431 { 00:30:51.431 "name": "pt3", 00:30:51.431 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:51.431 "is_configured": true, 00:30:51.431 "data_offset": 2048, 00:30:51.431 "data_size": 63488 00:30:51.431 }, 00:30:51.431 { 00:30:51.431 "name": null, 00:30:51.431 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:51.431 "is_configured": false, 00:30:51.431 "data_offset": 2048, 00:30:51.431 "data_size": 63488 00:30:51.431 } 00:30:51.431 ] 00:30:51.431 }' 00:30:51.431 20:29:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:51.431 20:29:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:52.012 [2024-10-01 20:29:47.036990] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:30:52.012 [2024-10-01 20:29:47.037085] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:52.012 [2024-10-01 20:29:47.037129] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:30:52.012 [2024-10-01 20:29:47.037149] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:52.012 [2024-10-01 20:29:47.037852] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:52.012 [2024-10-01 20:29:47.038048] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:30:52.012 [2024-10-01 20:29:47.038193] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:30:52.012 [2024-10-01 20:29:47.038234] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:30:52.012 [2024-10-01 20:29:47.038425] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:30:52.012 [2024-10-01 20:29:47.038444] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:30:52.012 [2024-10-01 20:29:47.038796] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:30:52.012 [2024-10-01 20:29:47.045675] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:30:52.012 [2024-10-01 20:29:47.045712] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:30:52.012 [2024-10-01 20:29:47.046089] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:52.012 pt4 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:52.012 "name": "raid_bdev1", 00:30:52.012 "uuid": "88b7386e-e410-4597-bbaa-4f19716a85a5", 00:30:52.012 "strip_size_kb": 64, 00:30:52.012 "state": "online", 00:30:52.012 "raid_level": "raid5f", 00:30:52.012 "superblock": true, 00:30:52.012 "num_base_bdevs": 4, 00:30:52.012 "num_base_bdevs_discovered": 3, 00:30:52.012 "num_base_bdevs_operational": 3, 00:30:52.012 "base_bdevs_list": [ 00:30:52.012 { 00:30:52.012 "name": null, 00:30:52.012 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:52.012 "is_configured": false, 00:30:52.012 "data_offset": 2048, 00:30:52.012 "data_size": 63488 00:30:52.012 }, 00:30:52.012 { 00:30:52.012 "name": "pt2", 00:30:52.012 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:52.012 "is_configured": true, 00:30:52.012 "data_offset": 2048, 00:30:52.012 "data_size": 63488 00:30:52.012 }, 00:30:52.012 { 00:30:52.012 "name": "pt3", 00:30:52.012 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:52.012 "is_configured": true, 00:30:52.012 "data_offset": 2048, 00:30:52.012 "data_size": 63488 00:30:52.012 }, 00:30:52.012 { 00:30:52.012 "name": "pt4", 00:30:52.012 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:52.012 "is_configured": true, 00:30:52.012 "data_offset": 2048, 00:30:52.012 "data_size": 63488 00:30:52.012 } 00:30:52.012 ] 00:30:52.012 }' 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:52.012 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:52.578 [2024-10-01 20:29:47.569662] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:30:52.578 [2024-10-01 20:29:47.569871] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:30:52.578 [2024-10-01 20:29:47.570013] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:52.578 [2024-10-01 20:29:47.570121] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:52.578 [2024-10-01 20:29:47.570148] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:52.578 [2024-10-01 20:29:47.641638] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:30:52.578 [2024-10-01 20:29:47.641748] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:52.578 [2024-10-01 20:29:47.641783] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:30:52.578 [2024-10-01 20:29:47.641809] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:52.578 [2024-10-01 20:29:47.644841] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:52.578 [2024-10-01 20:29:47.644898] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:30:52.578 [2024-10-01 20:29:47.645016] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:30:52.578 [2024-10-01 20:29:47.645099] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:30:52.578 [2024-10-01 20:29:47.645282] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:30:52.578 [2024-10-01 20:29:47.645322] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:30:52.578 [2024-10-01 20:29:47.645348] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:30:52.578 [2024-10-01 20:29:47.645446] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:30:52.578 [2024-10-01 20:29:47.645611] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:30:52.578 pt1 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:52.578 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:52.579 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:52.579 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:52.579 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:52.579 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:52.579 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:52.579 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:52.579 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:52.579 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:52.579 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:52.579 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:52.579 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:52.579 "name": "raid_bdev1", 00:30:52.579 "uuid": "88b7386e-e410-4597-bbaa-4f19716a85a5", 00:30:52.579 "strip_size_kb": 64, 00:30:52.579 "state": "configuring", 00:30:52.579 "raid_level": "raid5f", 00:30:52.579 "superblock": true, 00:30:52.579 "num_base_bdevs": 4, 00:30:52.579 "num_base_bdevs_discovered": 2, 00:30:52.579 "num_base_bdevs_operational": 3, 00:30:52.579 "base_bdevs_list": [ 00:30:52.579 { 00:30:52.579 "name": null, 00:30:52.579 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:52.579 "is_configured": false, 00:30:52.579 "data_offset": 2048, 00:30:52.579 "data_size": 63488 00:30:52.579 }, 00:30:52.579 { 00:30:52.579 "name": "pt2", 00:30:52.579 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:52.579 "is_configured": true, 00:30:52.579 "data_offset": 2048, 00:30:52.579 "data_size": 63488 00:30:52.579 }, 00:30:52.579 { 00:30:52.579 "name": "pt3", 00:30:52.579 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:52.579 "is_configured": true, 00:30:52.579 "data_offset": 2048, 00:30:52.579 "data_size": 63488 00:30:52.579 }, 00:30:52.579 { 00:30:52.579 "name": null, 00:30:52.579 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:52.579 "is_configured": false, 00:30:52.579 "data_offset": 2048, 00:30:52.579 "data_size": 63488 00:30:52.579 } 00:30:52.579 ] 00:30:52.579 }' 00:30:52.579 20:29:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:52.579 20:29:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:53.147 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:30:53.147 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:53.147 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:53.147 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:30:53.147 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:53.147 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:30:53.147 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:30:53.147 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:53.147 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:53.148 [2024-10-01 20:29:48.221896] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:30:53.148 [2024-10-01 20:29:48.222154] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:53.148 [2024-10-01 20:29:48.222210] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c680 00:30:53.148 [2024-10-01 20:29:48.222232] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:53.148 [2024-10-01 20:29:48.222964] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:53.148 [2024-10-01 20:29:48.222994] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:30:53.148 [2024-10-01 20:29:48.223118] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:30:53.148 [2024-10-01 20:29:48.223185] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:30:53.148 [2024-10-01 20:29:48.223360] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:30:53.148 [2024-10-01 20:29:48.223396] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:30:53.148 [2024-10-01 20:29:48.223760] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:30:53.148 [2024-10-01 20:29:48.231213] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:30:53.148 [2024-10-01 20:29:48.231263] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:30:53.148 [2024-10-01 20:29:48.231648] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:53.148 pt4 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:53.148 "name": "raid_bdev1", 00:30:53.148 "uuid": "88b7386e-e410-4597-bbaa-4f19716a85a5", 00:30:53.148 "strip_size_kb": 64, 00:30:53.148 "state": "online", 00:30:53.148 "raid_level": "raid5f", 00:30:53.148 "superblock": true, 00:30:53.148 "num_base_bdevs": 4, 00:30:53.148 "num_base_bdevs_discovered": 3, 00:30:53.148 "num_base_bdevs_operational": 3, 00:30:53.148 "base_bdevs_list": [ 00:30:53.148 { 00:30:53.148 "name": null, 00:30:53.148 "uuid": "00000000-0000-0000-0000-000000000000", 00:30:53.148 "is_configured": false, 00:30:53.148 "data_offset": 2048, 00:30:53.148 "data_size": 63488 00:30:53.148 }, 00:30:53.148 { 00:30:53.148 "name": "pt2", 00:30:53.148 "uuid": "00000000-0000-0000-0000-000000000002", 00:30:53.148 "is_configured": true, 00:30:53.148 "data_offset": 2048, 00:30:53.148 "data_size": 63488 00:30:53.148 }, 00:30:53.148 { 00:30:53.148 "name": "pt3", 00:30:53.148 "uuid": "00000000-0000-0000-0000-000000000003", 00:30:53.148 "is_configured": true, 00:30:53.148 "data_offset": 2048, 00:30:53.148 "data_size": 63488 00:30:53.148 }, 00:30:53.148 { 00:30:53.148 "name": "pt4", 00:30:53.148 "uuid": "00000000-0000-0000-0000-000000000004", 00:30:53.148 "is_configured": true, 00:30:53.148 "data_offset": 2048, 00:30:53.148 "data_size": 63488 00:30:53.148 } 00:30:53.148 ] 00:30:53.148 }' 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:53.148 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:30:53.715 [2024-10-01 20:29:48.835895] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 88b7386e-e410-4597-bbaa-4f19716a85a5 '!=' 88b7386e-e410-4597-bbaa-4f19716a85a5 ']' 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 85460 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 85460 ']' 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # kill -0 85460 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@955 -- # uname 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 85460 00:30:53.715 killing process with pid 85460 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 85460' 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@969 -- # kill 85460 00:30:53.715 [2024-10-01 20:29:48.917823] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:30:53.715 20:29:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@974 -- # wait 85460 00:30:53.715 [2024-10-01 20:29:48.917945] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:30:53.715 [2024-10-01 20:29:48.918052] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:30:53.715 [2024-10-01 20:29:48.918077] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:30:54.289 [2024-10-01 20:29:49.296548] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:30:56.831 20:29:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:30:56.831 00:30:56.831 real 0m11.045s 00:30:56.831 user 0m17.167s 00:30:56.831 sys 0m1.622s 00:30:56.831 ************************************ 00:30:56.831 END TEST raid5f_superblock_test 00:30:56.831 ************************************ 00:30:56.831 20:29:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:30:56.831 20:29:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:30:56.831 20:29:51 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:30:56.831 20:29:51 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 4 false false true 00:30:56.831 20:29:51 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:30:56.831 20:29:51 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:30:56.831 20:29:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:30:56.831 ************************************ 00:30:56.831 START TEST raid5f_rebuild_test 00:30:56.831 ************************************ 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid5f 4 false false true 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=85968 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 85968 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@831 -- # '[' -z 85968 ']' 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:56.831 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:30:56.831 20:29:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:56.831 I/O size of 3145728 is greater than zero copy threshold (65536). 00:30:56.831 Zero copy mechanism will not be used. 00:30:56.831 [2024-10-01 20:29:51.644027] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:30:56.831 [2024-10-01 20:29:51.644211] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85968 ] 00:30:56.831 [2024-10-01 20:29:51.822413] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:30:57.089 [2024-10-01 20:29:52.101276] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:30:57.349 [2024-10-01 20:29:52.347796] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:57.349 [2024-10-01 20:29:52.347866] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:30:57.608 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:30:57.608 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # return 0 00:30:57.608 20:29:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:30:57.608 20:29:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:30:57.608 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.608 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:57.970 BaseBdev1_malloc 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:57.970 [2024-10-01 20:29:52.881992] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:30:57.970 [2024-10-01 20:29:52.882125] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:57.970 [2024-10-01 20:29:52.882164] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:30:57.970 [2024-10-01 20:29:52.882189] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:57.970 [2024-10-01 20:29:52.885436] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:57.970 [2024-10-01 20:29:52.885490] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:30:57.970 BaseBdev1 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:57.970 BaseBdev2_malloc 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:57.970 [2024-10-01 20:29:52.947990] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:30:57.970 [2024-10-01 20:29:52.948102] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:57.970 [2024-10-01 20:29:52.948173] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:30:57.970 [2024-10-01 20:29:52.948208] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:57.970 [2024-10-01 20:29:52.951709] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:57.970 [2024-10-01 20:29:52.951793] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:30:57.970 BaseBdev2 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:57.970 BaseBdev3_malloc 00:30:57.970 20:29:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:57.970 [2024-10-01 20:29:53.008497] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:30:57.970 [2024-10-01 20:29:53.008577] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:57.970 [2024-10-01 20:29:53.008615] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:30:57.970 [2024-10-01 20:29:53.008637] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:57.970 [2024-10-01 20:29:53.011569] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:57.970 [2024-10-01 20:29:53.011626] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:30:57.970 BaseBdev3 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:57.970 BaseBdev4_malloc 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:57.970 [2024-10-01 20:29:53.064653] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:30:57.970 [2024-10-01 20:29:53.064778] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:57.970 [2024-10-01 20:29:53.064813] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:30:57.970 [2024-10-01 20:29:53.064833] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:57.970 [2024-10-01 20:29:53.067964] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:57.970 [2024-10-01 20:29:53.068022] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:30:57.970 BaseBdev4 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:57.970 spare_malloc 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:57.970 spare_delay 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:57.970 [2024-10-01 20:29:53.143378] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:30:57.970 [2024-10-01 20:29:53.143530] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:30:57.970 [2024-10-01 20:29:53.143574] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:30:57.970 [2024-10-01 20:29:53.143596] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:30:57.970 [2024-10-01 20:29:53.147471] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:30:57.970 [2024-10-01 20:29:53.147570] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:30:57.970 spare 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:57.970 [2024-10-01 20:29:53.156033] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:30:57.970 [2024-10-01 20:29:53.159036] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:30:57.970 [2024-10-01 20:29:53.159153] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:30:57.970 [2024-10-01 20:29:53.159255] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:30:57.970 [2024-10-01 20:29:53.159397] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:30:57.970 [2024-10-01 20:29:53.159420] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:30:57.970 [2024-10-01 20:29:53.159809] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:30:57.970 [2024-10-01 20:29:53.167213] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:30:57.970 [2024-10-01 20:29:53.167260] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:30:57.970 [2024-10-01 20:29:53.167600] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.970 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:30:57.971 "name": "raid_bdev1", 00:30:57.971 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:30:57.971 "strip_size_kb": 64, 00:30:57.971 "state": "online", 00:30:57.971 "raid_level": "raid5f", 00:30:57.971 "superblock": false, 00:30:57.971 "num_base_bdevs": 4, 00:30:57.971 "num_base_bdevs_discovered": 4, 00:30:57.971 "num_base_bdevs_operational": 4, 00:30:57.971 "base_bdevs_list": [ 00:30:57.971 { 00:30:57.971 "name": "BaseBdev1", 00:30:57.971 "uuid": "9ca16c3a-c7b0-5373-98a0-7f124ca49be7", 00:30:57.971 "is_configured": true, 00:30:57.971 "data_offset": 0, 00:30:57.971 "data_size": 65536 00:30:57.971 }, 00:30:57.971 { 00:30:57.971 "name": "BaseBdev2", 00:30:57.971 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:30:57.971 "is_configured": true, 00:30:57.971 "data_offset": 0, 00:30:57.971 "data_size": 65536 00:30:57.971 }, 00:30:57.971 { 00:30:57.971 "name": "BaseBdev3", 00:30:57.971 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:30:57.971 "is_configured": true, 00:30:57.971 "data_offset": 0, 00:30:57.971 "data_size": 65536 00:30:57.971 }, 00:30:57.971 { 00:30:57.971 "name": "BaseBdev4", 00:30:57.971 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:30:57.971 "is_configured": true, 00:30:57.971 "data_offset": 0, 00:30:57.971 "data_size": 65536 00:30:57.971 } 00:30:57.971 ] 00:30:57.971 }' 00:30:58.229 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:30:58.229 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:58.488 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:30:58.488 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:30:58.488 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.488 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:58.488 [2024-10-01 20:29:53.691776] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:30:58.488 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.488 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=196608 00:30:58.488 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:30:58.747 20:29:53 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:30:59.006 [2024-10-01 20:29:54.091876] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:30:59.006 /dev/nbd0 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:30:59.006 1+0 records in 00:30:59.006 1+0 records out 00:30:59.006 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000264878 s, 15.5 MB/s 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 192 00:30:59.006 20:29:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=512 oflag=direct 00:30:59.941 512+0 records in 00:30:59.942 512+0 records out 00:30:59.942 100663296 bytes (101 MB, 96 MiB) copied, 0.695525 s, 145 MB/s 00:30:59.942 20:29:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:30:59.942 20:29:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:30:59.942 20:29:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:30:59.942 20:29:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:30:59.942 20:29:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:30:59.942 20:29:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:30:59.942 20:29:54 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:30:59.942 [2024-10-01 20:29:55.154142] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:30:59.942 [2024-10-01 20:29:55.170051] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:59.942 20:29:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:00.200 20:29:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:00.200 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:00.200 "name": "raid_bdev1", 00:31:00.200 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:31:00.200 "strip_size_kb": 64, 00:31:00.200 "state": "online", 00:31:00.200 "raid_level": "raid5f", 00:31:00.200 "superblock": false, 00:31:00.200 "num_base_bdevs": 4, 00:31:00.200 "num_base_bdevs_discovered": 3, 00:31:00.200 "num_base_bdevs_operational": 3, 00:31:00.200 "base_bdevs_list": [ 00:31:00.200 { 00:31:00.200 "name": null, 00:31:00.200 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:00.200 "is_configured": false, 00:31:00.200 "data_offset": 0, 00:31:00.200 "data_size": 65536 00:31:00.200 }, 00:31:00.200 { 00:31:00.200 "name": "BaseBdev2", 00:31:00.200 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:31:00.200 "is_configured": true, 00:31:00.200 "data_offset": 0, 00:31:00.200 "data_size": 65536 00:31:00.200 }, 00:31:00.200 { 00:31:00.200 "name": "BaseBdev3", 00:31:00.200 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:31:00.200 "is_configured": true, 00:31:00.200 "data_offset": 0, 00:31:00.200 "data_size": 65536 00:31:00.200 }, 00:31:00.200 { 00:31:00.200 "name": "BaseBdev4", 00:31:00.200 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:31:00.200 "is_configured": true, 00:31:00.200 "data_offset": 0, 00:31:00.200 "data_size": 65536 00:31:00.200 } 00:31:00.200 ] 00:31:00.200 }' 00:31:00.200 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:00.200 20:29:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:00.460 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:31:00.460 20:29:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:00.460 20:29:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:00.460 [2024-10-01 20:29:55.682222] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:00.460 [2024-10-01 20:29:55.699356] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b750 00:31:00.460 20:29:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:00.460 20:29:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:31:00.460 [2024-10-01 20:29:55.710615] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:31:01.836 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:01.836 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:01.836 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:01.836 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:01.836 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:01.836 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:01.836 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:01.836 20:29:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:01.836 20:29:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:01.836 20:29:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:01.836 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:01.836 "name": "raid_bdev1", 00:31:01.836 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:31:01.836 "strip_size_kb": 64, 00:31:01.836 "state": "online", 00:31:01.836 "raid_level": "raid5f", 00:31:01.836 "superblock": false, 00:31:01.836 "num_base_bdevs": 4, 00:31:01.836 "num_base_bdevs_discovered": 4, 00:31:01.836 "num_base_bdevs_operational": 4, 00:31:01.836 "process": { 00:31:01.836 "type": "rebuild", 00:31:01.836 "target": "spare", 00:31:01.836 "progress": { 00:31:01.836 "blocks": 17280, 00:31:01.836 "percent": 8 00:31:01.836 } 00:31:01.836 }, 00:31:01.836 "base_bdevs_list": [ 00:31:01.836 { 00:31:01.836 "name": "spare", 00:31:01.836 "uuid": "4201956e-d67b-53d3-be35-4882d0c19c25", 00:31:01.836 "is_configured": true, 00:31:01.836 "data_offset": 0, 00:31:01.836 "data_size": 65536 00:31:01.836 }, 00:31:01.836 { 00:31:01.836 "name": "BaseBdev2", 00:31:01.837 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:31:01.837 "is_configured": true, 00:31:01.837 "data_offset": 0, 00:31:01.837 "data_size": 65536 00:31:01.837 }, 00:31:01.837 { 00:31:01.837 "name": "BaseBdev3", 00:31:01.837 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:31:01.837 "is_configured": true, 00:31:01.837 "data_offset": 0, 00:31:01.837 "data_size": 65536 00:31:01.837 }, 00:31:01.837 { 00:31:01.837 "name": "BaseBdev4", 00:31:01.837 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:31:01.837 "is_configured": true, 00:31:01.837 "data_offset": 0, 00:31:01.837 "data_size": 65536 00:31:01.837 } 00:31:01.837 ] 00:31:01.837 }' 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:01.837 [2024-10-01 20:29:56.868732] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:01.837 [2024-10-01 20:29:56.925058] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:31:01.837 [2024-10-01 20:29:56.925185] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:01.837 [2024-10-01 20:29:56.925266] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:01.837 [2024-10-01 20:29:56.925283] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:01.837 20:29:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:01.837 20:29:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:01.837 "name": "raid_bdev1", 00:31:01.837 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:31:01.837 "strip_size_kb": 64, 00:31:01.837 "state": "online", 00:31:01.837 "raid_level": "raid5f", 00:31:01.837 "superblock": false, 00:31:01.837 "num_base_bdevs": 4, 00:31:01.837 "num_base_bdevs_discovered": 3, 00:31:01.837 "num_base_bdevs_operational": 3, 00:31:01.837 "base_bdevs_list": [ 00:31:01.837 { 00:31:01.837 "name": null, 00:31:01.837 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:01.837 "is_configured": false, 00:31:01.837 "data_offset": 0, 00:31:01.837 "data_size": 65536 00:31:01.837 }, 00:31:01.837 { 00:31:01.837 "name": "BaseBdev2", 00:31:01.837 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:31:01.837 "is_configured": true, 00:31:01.837 "data_offset": 0, 00:31:01.837 "data_size": 65536 00:31:01.837 }, 00:31:01.837 { 00:31:01.837 "name": "BaseBdev3", 00:31:01.837 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:31:01.837 "is_configured": true, 00:31:01.837 "data_offset": 0, 00:31:01.837 "data_size": 65536 00:31:01.837 }, 00:31:01.837 { 00:31:01.837 "name": "BaseBdev4", 00:31:01.837 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:31:01.837 "is_configured": true, 00:31:01.837 "data_offset": 0, 00:31:01.837 "data_size": 65536 00:31:01.837 } 00:31:01.837 ] 00:31:01.837 }' 00:31:01.837 20:29:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:01.837 20:29:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:02.404 20:29:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:02.404 20:29:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:02.404 20:29:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:02.404 20:29:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:02.404 20:29:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:02.404 20:29:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:02.404 20:29:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:02.404 20:29:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:02.404 20:29:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:02.404 20:29:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:02.404 20:29:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:02.404 "name": "raid_bdev1", 00:31:02.404 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:31:02.404 "strip_size_kb": 64, 00:31:02.404 "state": "online", 00:31:02.404 "raid_level": "raid5f", 00:31:02.404 "superblock": false, 00:31:02.404 "num_base_bdevs": 4, 00:31:02.404 "num_base_bdevs_discovered": 3, 00:31:02.404 "num_base_bdevs_operational": 3, 00:31:02.404 "base_bdevs_list": [ 00:31:02.404 { 00:31:02.404 "name": null, 00:31:02.404 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:02.404 "is_configured": false, 00:31:02.404 "data_offset": 0, 00:31:02.404 "data_size": 65536 00:31:02.404 }, 00:31:02.404 { 00:31:02.404 "name": "BaseBdev2", 00:31:02.404 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:31:02.404 "is_configured": true, 00:31:02.404 "data_offset": 0, 00:31:02.404 "data_size": 65536 00:31:02.404 }, 00:31:02.404 { 00:31:02.404 "name": "BaseBdev3", 00:31:02.404 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:31:02.404 "is_configured": true, 00:31:02.404 "data_offset": 0, 00:31:02.404 "data_size": 65536 00:31:02.404 }, 00:31:02.404 { 00:31:02.404 "name": "BaseBdev4", 00:31:02.404 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:31:02.404 "is_configured": true, 00:31:02.404 "data_offset": 0, 00:31:02.404 "data_size": 65536 00:31:02.404 } 00:31:02.404 ] 00:31:02.404 }' 00:31:02.404 20:29:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:02.404 20:29:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:02.404 20:29:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:02.663 20:29:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:02.663 20:29:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:31:02.663 20:29:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:02.663 20:29:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:02.663 [2024-10-01 20:29:57.670980] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:02.663 [2024-10-01 20:29:57.686125] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002b820 00:31:02.663 20:29:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:02.663 20:29:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:31:02.663 [2024-10-01 20:29:57.697283] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:31:03.599 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:03.599 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:03.599 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:03.599 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:03.599 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:03.599 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:03.599 20:29:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:03.599 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:03.599 20:29:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:03.599 20:29:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:03.599 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:03.599 "name": "raid_bdev1", 00:31:03.599 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:31:03.599 "strip_size_kb": 64, 00:31:03.599 "state": "online", 00:31:03.599 "raid_level": "raid5f", 00:31:03.599 "superblock": false, 00:31:03.599 "num_base_bdevs": 4, 00:31:03.599 "num_base_bdevs_discovered": 4, 00:31:03.599 "num_base_bdevs_operational": 4, 00:31:03.599 "process": { 00:31:03.599 "type": "rebuild", 00:31:03.599 "target": "spare", 00:31:03.599 "progress": { 00:31:03.599 "blocks": 17280, 00:31:03.599 "percent": 8 00:31:03.599 } 00:31:03.599 }, 00:31:03.599 "base_bdevs_list": [ 00:31:03.599 { 00:31:03.599 "name": "spare", 00:31:03.599 "uuid": "4201956e-d67b-53d3-be35-4882d0c19c25", 00:31:03.599 "is_configured": true, 00:31:03.599 "data_offset": 0, 00:31:03.599 "data_size": 65536 00:31:03.599 }, 00:31:03.599 { 00:31:03.599 "name": "BaseBdev2", 00:31:03.599 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:31:03.599 "is_configured": true, 00:31:03.599 "data_offset": 0, 00:31:03.599 "data_size": 65536 00:31:03.599 }, 00:31:03.599 { 00:31:03.599 "name": "BaseBdev3", 00:31:03.599 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:31:03.599 "is_configured": true, 00:31:03.599 "data_offset": 0, 00:31:03.599 "data_size": 65536 00:31:03.599 }, 00:31:03.599 { 00:31:03.599 "name": "BaseBdev4", 00:31:03.599 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:31:03.599 "is_configured": true, 00:31:03.599 "data_offset": 0, 00:31:03.599 "data_size": 65536 00:31:03.599 } 00:31:03.599 ] 00:31:03.599 }' 00:31:03.599 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:03.599 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:03.599 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=737 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:03.857 "name": "raid_bdev1", 00:31:03.857 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:31:03.857 "strip_size_kb": 64, 00:31:03.857 "state": "online", 00:31:03.857 "raid_level": "raid5f", 00:31:03.857 "superblock": false, 00:31:03.857 "num_base_bdevs": 4, 00:31:03.857 "num_base_bdevs_discovered": 4, 00:31:03.857 "num_base_bdevs_operational": 4, 00:31:03.857 "process": { 00:31:03.857 "type": "rebuild", 00:31:03.857 "target": "spare", 00:31:03.857 "progress": { 00:31:03.857 "blocks": 21120, 00:31:03.857 "percent": 10 00:31:03.857 } 00:31:03.857 }, 00:31:03.857 "base_bdevs_list": [ 00:31:03.857 { 00:31:03.857 "name": "spare", 00:31:03.857 "uuid": "4201956e-d67b-53d3-be35-4882d0c19c25", 00:31:03.857 "is_configured": true, 00:31:03.857 "data_offset": 0, 00:31:03.857 "data_size": 65536 00:31:03.857 }, 00:31:03.857 { 00:31:03.857 "name": "BaseBdev2", 00:31:03.857 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:31:03.857 "is_configured": true, 00:31:03.857 "data_offset": 0, 00:31:03.857 "data_size": 65536 00:31:03.857 }, 00:31:03.857 { 00:31:03.857 "name": "BaseBdev3", 00:31:03.857 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:31:03.857 "is_configured": true, 00:31:03.857 "data_offset": 0, 00:31:03.857 "data_size": 65536 00:31:03.857 }, 00:31:03.857 { 00:31:03.857 "name": "BaseBdev4", 00:31:03.857 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:31:03.857 "is_configured": true, 00:31:03.857 "data_offset": 0, 00:31:03.857 "data_size": 65536 00:31:03.857 } 00:31:03.857 ] 00:31:03.857 }' 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:03.857 20:29:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:03.857 20:29:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:03.857 20:29:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:04.794 20:30:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:04.794 20:30:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:04.794 20:30:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:04.794 20:30:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:04.794 20:30:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:04.794 20:30:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:05.053 20:30:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:05.053 20:30:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:05.053 20:30:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:05.053 20:30:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:05.053 20:30:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:05.053 20:30:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:05.053 "name": "raid_bdev1", 00:31:05.053 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:31:05.053 "strip_size_kb": 64, 00:31:05.053 "state": "online", 00:31:05.053 "raid_level": "raid5f", 00:31:05.053 "superblock": false, 00:31:05.053 "num_base_bdevs": 4, 00:31:05.053 "num_base_bdevs_discovered": 4, 00:31:05.053 "num_base_bdevs_operational": 4, 00:31:05.053 "process": { 00:31:05.053 "type": "rebuild", 00:31:05.053 "target": "spare", 00:31:05.053 "progress": { 00:31:05.053 "blocks": 44160, 00:31:05.053 "percent": 22 00:31:05.053 } 00:31:05.053 }, 00:31:05.053 "base_bdevs_list": [ 00:31:05.053 { 00:31:05.053 "name": "spare", 00:31:05.053 "uuid": "4201956e-d67b-53d3-be35-4882d0c19c25", 00:31:05.053 "is_configured": true, 00:31:05.053 "data_offset": 0, 00:31:05.053 "data_size": 65536 00:31:05.053 }, 00:31:05.053 { 00:31:05.053 "name": "BaseBdev2", 00:31:05.053 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:31:05.053 "is_configured": true, 00:31:05.053 "data_offset": 0, 00:31:05.053 "data_size": 65536 00:31:05.053 }, 00:31:05.053 { 00:31:05.053 "name": "BaseBdev3", 00:31:05.053 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:31:05.053 "is_configured": true, 00:31:05.053 "data_offset": 0, 00:31:05.053 "data_size": 65536 00:31:05.053 }, 00:31:05.053 { 00:31:05.053 "name": "BaseBdev4", 00:31:05.053 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:31:05.053 "is_configured": true, 00:31:05.053 "data_offset": 0, 00:31:05.053 "data_size": 65536 00:31:05.053 } 00:31:05.053 ] 00:31:05.053 }' 00:31:05.053 20:30:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:05.053 20:30:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:05.053 20:30:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:05.053 20:30:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:05.053 20:30:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:05.987 20:30:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:05.987 20:30:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:05.987 20:30:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:05.987 20:30:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:05.987 20:30:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:05.987 20:30:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:05.987 20:30:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:05.987 20:30:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:05.987 20:30:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:05.987 20:30:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:05.987 20:30:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:06.245 20:30:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:06.245 "name": "raid_bdev1", 00:31:06.245 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:31:06.245 "strip_size_kb": 64, 00:31:06.245 "state": "online", 00:31:06.245 "raid_level": "raid5f", 00:31:06.245 "superblock": false, 00:31:06.245 "num_base_bdevs": 4, 00:31:06.245 "num_base_bdevs_discovered": 4, 00:31:06.245 "num_base_bdevs_operational": 4, 00:31:06.245 "process": { 00:31:06.245 "type": "rebuild", 00:31:06.245 "target": "spare", 00:31:06.245 "progress": { 00:31:06.245 "blocks": 65280, 00:31:06.245 "percent": 33 00:31:06.245 } 00:31:06.245 }, 00:31:06.245 "base_bdevs_list": [ 00:31:06.245 { 00:31:06.245 "name": "spare", 00:31:06.245 "uuid": "4201956e-d67b-53d3-be35-4882d0c19c25", 00:31:06.245 "is_configured": true, 00:31:06.245 "data_offset": 0, 00:31:06.245 "data_size": 65536 00:31:06.245 }, 00:31:06.245 { 00:31:06.245 "name": "BaseBdev2", 00:31:06.245 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:31:06.245 "is_configured": true, 00:31:06.245 "data_offset": 0, 00:31:06.245 "data_size": 65536 00:31:06.245 }, 00:31:06.245 { 00:31:06.245 "name": "BaseBdev3", 00:31:06.245 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:31:06.245 "is_configured": true, 00:31:06.245 "data_offset": 0, 00:31:06.245 "data_size": 65536 00:31:06.245 }, 00:31:06.245 { 00:31:06.245 "name": "BaseBdev4", 00:31:06.245 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:31:06.245 "is_configured": true, 00:31:06.245 "data_offset": 0, 00:31:06.245 "data_size": 65536 00:31:06.245 } 00:31:06.245 ] 00:31:06.245 }' 00:31:06.245 20:30:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:06.245 20:30:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:06.245 20:30:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:06.245 20:30:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:06.245 20:30:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:07.182 20:30:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:07.182 20:30:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:07.182 20:30:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:07.182 20:30:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:07.182 20:30:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:07.182 20:30:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:07.182 20:30:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:07.182 20:30:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:07.182 20:30:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:07.182 20:30:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:07.182 20:30:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:07.182 20:30:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:07.182 "name": "raid_bdev1", 00:31:07.182 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:31:07.182 "strip_size_kb": 64, 00:31:07.182 "state": "online", 00:31:07.182 "raid_level": "raid5f", 00:31:07.182 "superblock": false, 00:31:07.182 "num_base_bdevs": 4, 00:31:07.182 "num_base_bdevs_discovered": 4, 00:31:07.182 "num_base_bdevs_operational": 4, 00:31:07.182 "process": { 00:31:07.182 "type": "rebuild", 00:31:07.182 "target": "spare", 00:31:07.182 "progress": { 00:31:07.182 "blocks": 88320, 00:31:07.182 "percent": 44 00:31:07.182 } 00:31:07.182 }, 00:31:07.182 "base_bdevs_list": [ 00:31:07.182 { 00:31:07.182 "name": "spare", 00:31:07.182 "uuid": "4201956e-d67b-53d3-be35-4882d0c19c25", 00:31:07.182 "is_configured": true, 00:31:07.182 "data_offset": 0, 00:31:07.182 "data_size": 65536 00:31:07.182 }, 00:31:07.182 { 00:31:07.182 "name": "BaseBdev2", 00:31:07.182 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:31:07.182 "is_configured": true, 00:31:07.182 "data_offset": 0, 00:31:07.182 "data_size": 65536 00:31:07.182 }, 00:31:07.182 { 00:31:07.182 "name": "BaseBdev3", 00:31:07.182 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:31:07.182 "is_configured": true, 00:31:07.182 "data_offset": 0, 00:31:07.182 "data_size": 65536 00:31:07.182 }, 00:31:07.182 { 00:31:07.182 "name": "BaseBdev4", 00:31:07.182 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:31:07.182 "is_configured": true, 00:31:07.182 "data_offset": 0, 00:31:07.182 "data_size": 65536 00:31:07.182 } 00:31:07.182 ] 00:31:07.182 }' 00:31:07.182 20:30:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:07.441 20:30:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:07.441 20:30:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:07.441 20:30:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:07.441 20:30:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:08.412 20:30:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:08.412 20:30:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:08.412 20:30:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:08.412 20:30:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:08.412 20:30:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:08.412 20:30:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:08.412 20:30:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:08.412 20:30:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:08.412 20:30:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:08.412 20:30:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:08.412 20:30:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:08.412 20:30:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:08.412 "name": "raid_bdev1", 00:31:08.412 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:31:08.412 "strip_size_kb": 64, 00:31:08.412 "state": "online", 00:31:08.412 "raid_level": "raid5f", 00:31:08.412 "superblock": false, 00:31:08.412 "num_base_bdevs": 4, 00:31:08.412 "num_base_bdevs_discovered": 4, 00:31:08.412 "num_base_bdevs_operational": 4, 00:31:08.412 "process": { 00:31:08.412 "type": "rebuild", 00:31:08.412 "target": "spare", 00:31:08.412 "progress": { 00:31:08.412 "blocks": 109440, 00:31:08.412 "percent": 55 00:31:08.412 } 00:31:08.412 }, 00:31:08.412 "base_bdevs_list": [ 00:31:08.412 { 00:31:08.412 "name": "spare", 00:31:08.412 "uuid": "4201956e-d67b-53d3-be35-4882d0c19c25", 00:31:08.412 "is_configured": true, 00:31:08.412 "data_offset": 0, 00:31:08.412 "data_size": 65536 00:31:08.412 }, 00:31:08.412 { 00:31:08.412 "name": "BaseBdev2", 00:31:08.412 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:31:08.412 "is_configured": true, 00:31:08.412 "data_offset": 0, 00:31:08.412 "data_size": 65536 00:31:08.412 }, 00:31:08.412 { 00:31:08.412 "name": "BaseBdev3", 00:31:08.412 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:31:08.412 "is_configured": true, 00:31:08.412 "data_offset": 0, 00:31:08.412 "data_size": 65536 00:31:08.412 }, 00:31:08.412 { 00:31:08.412 "name": "BaseBdev4", 00:31:08.412 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:31:08.412 "is_configured": true, 00:31:08.412 "data_offset": 0, 00:31:08.412 "data_size": 65536 00:31:08.412 } 00:31:08.412 ] 00:31:08.412 }' 00:31:08.412 20:30:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:08.412 20:30:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:08.412 20:30:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:08.670 20:30:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:08.670 20:30:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:09.605 20:30:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:09.605 20:30:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:09.605 20:30:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:09.605 20:30:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:09.605 20:30:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:09.605 20:30:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:09.605 20:30:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:09.605 20:30:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.605 20:30:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:09.605 20:30:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:09.605 20:30:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.605 20:30:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:09.605 "name": "raid_bdev1", 00:31:09.605 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:31:09.605 "strip_size_kb": 64, 00:31:09.605 "state": "online", 00:31:09.605 "raid_level": "raid5f", 00:31:09.605 "superblock": false, 00:31:09.605 "num_base_bdevs": 4, 00:31:09.605 "num_base_bdevs_discovered": 4, 00:31:09.605 "num_base_bdevs_operational": 4, 00:31:09.605 "process": { 00:31:09.605 "type": "rebuild", 00:31:09.605 "target": "spare", 00:31:09.605 "progress": { 00:31:09.605 "blocks": 132480, 00:31:09.605 "percent": 67 00:31:09.605 } 00:31:09.605 }, 00:31:09.605 "base_bdevs_list": [ 00:31:09.605 { 00:31:09.605 "name": "spare", 00:31:09.605 "uuid": "4201956e-d67b-53d3-be35-4882d0c19c25", 00:31:09.605 "is_configured": true, 00:31:09.605 "data_offset": 0, 00:31:09.605 "data_size": 65536 00:31:09.605 }, 00:31:09.605 { 00:31:09.605 "name": "BaseBdev2", 00:31:09.605 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:31:09.605 "is_configured": true, 00:31:09.605 "data_offset": 0, 00:31:09.605 "data_size": 65536 00:31:09.605 }, 00:31:09.605 { 00:31:09.605 "name": "BaseBdev3", 00:31:09.605 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:31:09.605 "is_configured": true, 00:31:09.605 "data_offset": 0, 00:31:09.605 "data_size": 65536 00:31:09.606 }, 00:31:09.606 { 00:31:09.606 "name": "BaseBdev4", 00:31:09.606 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:31:09.606 "is_configured": true, 00:31:09.606 "data_offset": 0, 00:31:09.606 "data_size": 65536 00:31:09.606 } 00:31:09.606 ] 00:31:09.606 }' 00:31:09.606 20:30:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:09.606 20:30:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:09.606 20:30:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:09.864 20:30:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:09.864 20:30:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:10.799 20:30:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:10.799 20:30:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:10.799 20:30:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:10.799 20:30:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:10.799 20:30:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:10.799 20:30:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:10.799 20:30:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:10.799 20:30:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:10.799 20:30:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.799 20:30:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:10.799 20:30:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.799 20:30:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:10.799 "name": "raid_bdev1", 00:31:10.799 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:31:10.799 "strip_size_kb": 64, 00:31:10.799 "state": "online", 00:31:10.799 "raid_level": "raid5f", 00:31:10.799 "superblock": false, 00:31:10.799 "num_base_bdevs": 4, 00:31:10.799 "num_base_bdevs_discovered": 4, 00:31:10.799 "num_base_bdevs_operational": 4, 00:31:10.799 "process": { 00:31:10.799 "type": "rebuild", 00:31:10.799 "target": "spare", 00:31:10.799 "progress": { 00:31:10.799 "blocks": 155520, 00:31:10.799 "percent": 79 00:31:10.799 } 00:31:10.799 }, 00:31:10.799 "base_bdevs_list": [ 00:31:10.799 { 00:31:10.799 "name": "spare", 00:31:10.799 "uuid": "4201956e-d67b-53d3-be35-4882d0c19c25", 00:31:10.799 "is_configured": true, 00:31:10.799 "data_offset": 0, 00:31:10.799 "data_size": 65536 00:31:10.799 }, 00:31:10.799 { 00:31:10.799 "name": "BaseBdev2", 00:31:10.799 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:31:10.799 "is_configured": true, 00:31:10.799 "data_offset": 0, 00:31:10.799 "data_size": 65536 00:31:10.799 }, 00:31:10.799 { 00:31:10.799 "name": "BaseBdev3", 00:31:10.799 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:31:10.799 "is_configured": true, 00:31:10.799 "data_offset": 0, 00:31:10.799 "data_size": 65536 00:31:10.799 }, 00:31:10.799 { 00:31:10.799 "name": "BaseBdev4", 00:31:10.799 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:31:10.799 "is_configured": true, 00:31:10.799 "data_offset": 0, 00:31:10.799 "data_size": 65536 00:31:10.799 } 00:31:10.799 ] 00:31:10.799 }' 00:31:10.799 20:30:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:10.799 20:30:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:10.799 20:30:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:11.058 20:30:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:11.058 20:30:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:11.992 "name": "raid_bdev1", 00:31:11.992 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:31:11.992 "strip_size_kb": 64, 00:31:11.992 "state": "online", 00:31:11.992 "raid_level": "raid5f", 00:31:11.992 "superblock": false, 00:31:11.992 "num_base_bdevs": 4, 00:31:11.992 "num_base_bdevs_discovered": 4, 00:31:11.992 "num_base_bdevs_operational": 4, 00:31:11.992 "process": { 00:31:11.992 "type": "rebuild", 00:31:11.992 "target": "spare", 00:31:11.992 "progress": { 00:31:11.992 "blocks": 176640, 00:31:11.992 "percent": 89 00:31:11.992 } 00:31:11.992 }, 00:31:11.992 "base_bdevs_list": [ 00:31:11.992 { 00:31:11.992 "name": "spare", 00:31:11.992 "uuid": "4201956e-d67b-53d3-be35-4882d0c19c25", 00:31:11.992 "is_configured": true, 00:31:11.992 "data_offset": 0, 00:31:11.992 "data_size": 65536 00:31:11.992 }, 00:31:11.992 { 00:31:11.992 "name": "BaseBdev2", 00:31:11.992 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:31:11.992 "is_configured": true, 00:31:11.992 "data_offset": 0, 00:31:11.992 "data_size": 65536 00:31:11.992 }, 00:31:11.992 { 00:31:11.992 "name": "BaseBdev3", 00:31:11.992 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:31:11.992 "is_configured": true, 00:31:11.992 "data_offset": 0, 00:31:11.992 "data_size": 65536 00:31:11.992 }, 00:31:11.992 { 00:31:11.992 "name": "BaseBdev4", 00:31:11.992 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:31:11.992 "is_configured": true, 00:31:11.992 "data_offset": 0, 00:31:11.992 "data_size": 65536 00:31:11.992 } 00:31:11.992 ] 00:31:11.992 }' 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:11.992 20:30:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:12.926 [2024-10-01 20:30:08.109475] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:31:12.926 [2024-10-01 20:30:08.109621] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:31:12.926 [2024-10-01 20:30:08.109745] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:13.184 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:13.184 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:13.184 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:13.184 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:13.184 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:13.184 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:13.184 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:13.184 20:30:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:13.184 20:30:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:13.184 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:13.184 20:30:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:13.184 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:13.184 "name": "raid_bdev1", 00:31:13.184 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:31:13.185 "strip_size_kb": 64, 00:31:13.185 "state": "online", 00:31:13.185 "raid_level": "raid5f", 00:31:13.185 "superblock": false, 00:31:13.185 "num_base_bdevs": 4, 00:31:13.185 "num_base_bdevs_discovered": 4, 00:31:13.185 "num_base_bdevs_operational": 4, 00:31:13.185 "base_bdevs_list": [ 00:31:13.185 { 00:31:13.185 "name": "spare", 00:31:13.185 "uuid": "4201956e-d67b-53d3-be35-4882d0c19c25", 00:31:13.185 "is_configured": true, 00:31:13.185 "data_offset": 0, 00:31:13.185 "data_size": 65536 00:31:13.185 }, 00:31:13.185 { 00:31:13.185 "name": "BaseBdev2", 00:31:13.185 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:31:13.185 "is_configured": true, 00:31:13.185 "data_offset": 0, 00:31:13.185 "data_size": 65536 00:31:13.185 }, 00:31:13.185 { 00:31:13.185 "name": "BaseBdev3", 00:31:13.185 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:31:13.185 "is_configured": true, 00:31:13.185 "data_offset": 0, 00:31:13.185 "data_size": 65536 00:31:13.185 }, 00:31:13.185 { 00:31:13.185 "name": "BaseBdev4", 00:31:13.185 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:31:13.185 "is_configured": true, 00:31:13.185 "data_offset": 0, 00:31:13.185 "data_size": 65536 00:31:13.185 } 00:31:13.185 ] 00:31:13.185 }' 00:31:13.185 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:13.185 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:31:13.185 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:13.185 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:31:13.185 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:31:13.185 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:13.185 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:13.185 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:13.185 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:13.185 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:13.185 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:13.185 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:13.185 20:30:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:13.185 20:30:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:13.185 20:30:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:13.443 "name": "raid_bdev1", 00:31:13.443 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:31:13.443 "strip_size_kb": 64, 00:31:13.443 "state": "online", 00:31:13.443 "raid_level": "raid5f", 00:31:13.443 "superblock": false, 00:31:13.443 "num_base_bdevs": 4, 00:31:13.443 "num_base_bdevs_discovered": 4, 00:31:13.443 "num_base_bdevs_operational": 4, 00:31:13.443 "base_bdevs_list": [ 00:31:13.443 { 00:31:13.443 "name": "spare", 00:31:13.443 "uuid": "4201956e-d67b-53d3-be35-4882d0c19c25", 00:31:13.443 "is_configured": true, 00:31:13.443 "data_offset": 0, 00:31:13.443 "data_size": 65536 00:31:13.443 }, 00:31:13.443 { 00:31:13.443 "name": "BaseBdev2", 00:31:13.443 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:31:13.443 "is_configured": true, 00:31:13.443 "data_offset": 0, 00:31:13.443 "data_size": 65536 00:31:13.443 }, 00:31:13.443 { 00:31:13.443 "name": "BaseBdev3", 00:31:13.443 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:31:13.443 "is_configured": true, 00:31:13.443 "data_offset": 0, 00:31:13.443 "data_size": 65536 00:31:13.443 }, 00:31:13.443 { 00:31:13.443 "name": "BaseBdev4", 00:31:13.443 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:31:13.443 "is_configured": true, 00:31:13.443 "data_offset": 0, 00:31:13.443 "data_size": 65536 00:31:13.443 } 00:31:13.443 ] 00:31:13.443 }' 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:13.443 "name": "raid_bdev1", 00:31:13.443 "uuid": "ce86e39c-a15f-49aa-b64f-3d25005bd50d", 00:31:13.443 "strip_size_kb": 64, 00:31:13.443 "state": "online", 00:31:13.443 "raid_level": "raid5f", 00:31:13.443 "superblock": false, 00:31:13.443 "num_base_bdevs": 4, 00:31:13.443 "num_base_bdevs_discovered": 4, 00:31:13.443 "num_base_bdevs_operational": 4, 00:31:13.443 "base_bdevs_list": [ 00:31:13.443 { 00:31:13.443 "name": "spare", 00:31:13.443 "uuid": "4201956e-d67b-53d3-be35-4882d0c19c25", 00:31:13.443 "is_configured": true, 00:31:13.443 "data_offset": 0, 00:31:13.443 "data_size": 65536 00:31:13.443 }, 00:31:13.443 { 00:31:13.443 "name": "BaseBdev2", 00:31:13.443 "uuid": "7539aa37-ed85-5a71-9914-73bf6cd6a042", 00:31:13.443 "is_configured": true, 00:31:13.443 "data_offset": 0, 00:31:13.443 "data_size": 65536 00:31:13.443 }, 00:31:13.443 { 00:31:13.443 "name": "BaseBdev3", 00:31:13.443 "uuid": "bfe7d057-fe68-5642-92fa-6cc9e1991800", 00:31:13.443 "is_configured": true, 00:31:13.443 "data_offset": 0, 00:31:13.443 "data_size": 65536 00:31:13.443 }, 00:31:13.443 { 00:31:13.443 "name": "BaseBdev4", 00:31:13.443 "uuid": "2543cd04-129c-5a99-b3e5-705429203227", 00:31:13.443 "is_configured": true, 00:31:13.443 "data_offset": 0, 00:31:13.443 "data_size": 65536 00:31:13.443 } 00:31:13.443 ] 00:31:13.443 }' 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:13.443 20:30:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:14.008 [2024-10-01 20:30:09.073267] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:31:14.008 [2024-10-01 20:30:09.073325] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:31:14.008 [2024-10-01 20:30:09.073486] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:31:14.008 [2024-10-01 20:30:09.073612] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:31:14.008 [2024-10-01 20:30:09.073631] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:31:14.008 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:31:14.266 /dev/nbd0 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:14.266 1+0 records in 00:31:14.266 1+0 records out 00:31:14.266 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000308279 s, 13.3 MB/s 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:31:14.266 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:31:14.524 /dev/nbd1 00:31:14.524 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:31:14.524 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:31:14.524 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:31:14.524 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:31:14.524 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:31:14.524 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:31:14.524 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:31:14.524 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:31:14.524 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:31:14.524 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:31:14.524 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:14.524 1+0 records in 00:31:14.524 1+0 records out 00:31:14.524 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000342258 s, 12.0 MB/s 00:31:14.524 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:14.524 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:31:14.524 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:14.524 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:31:14.524 20:30:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:31:14.525 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:14.525 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:31:14.525 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:31:14.782 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:31:14.782 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:31:14.782 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:31:14.782 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:31:14.782 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:31:14.782 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:14.782 20:30:09 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:31:15.041 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:31:15.041 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:31:15.041 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:31:15.041 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:15.041 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:15.041 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:31:15.041 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:31:15.041 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:31:15.041 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:15.041 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 85968 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@950 -- # '[' -z 85968 ']' 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # kill -0 85968 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@955 -- # uname 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 85968 00:31:15.635 killing process with pid 85968 00:31:15.635 Received shutdown signal, test time was about 60.000000 seconds 00:31:15.635 00:31:15.635 Latency(us) 00:31:15.635 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:15.635 =================================================================================================================== 00:31:15.635 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 85968' 00:31:15.635 20:30:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@969 -- # kill 85968 00:31:15.636 20:30:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@974 -- # wait 85968 00:31:15.636 [2024-10-01 20:30:10.625421] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:31:15.893 [2024-10-01 20:30:11.145833] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:31:18.422 ************************************ 00:31:18.422 END TEST raid5f_rebuild_test 00:31:18.422 ************************************ 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:31:18.422 00:31:18.422 real 0m21.741s 00:31:18.422 user 0m26.505s 00:31:18.422 sys 0m2.678s 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:31:18.422 20:30:13 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 4 true false true 00:31:18.422 20:30:13 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:31:18.422 20:30:13 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:31:18.422 20:30:13 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:31:18.422 ************************************ 00:31:18.422 START TEST raid5f_rebuild_test_sb 00:31:18.422 ************************************ 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid5f 4 true false true 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=86488 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 86488 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@831 -- # '[' -z 86488 ']' 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:31:18.422 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:31:18.422 20:30:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:18.422 [2024-10-01 20:30:13.500093] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:31:18.422 I/O size of 3145728 is greater than zero copy threshold (65536). 00:31:18.422 Zero copy mechanism will not be used. 00:31:18.422 [2024-10-01 20:30:13.500329] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86488 ] 00:31:18.680 [2024-10-01 20:30:13.682029] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:18.938 [2024-10-01 20:30:13.948568] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:31:19.195 [2024-10-01 20:30:14.201789] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:31:19.196 [2024-10-01 20:30:14.201915] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:31:19.454 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:31:19.454 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # return 0 00:31:19.454 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:31:19.454 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:31:19.454 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.454 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:19.712 BaseBdev1_malloc 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:19.712 [2024-10-01 20:30:14.746575] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:31:19.712 [2024-10-01 20:30:14.746662] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:19.712 [2024-10-01 20:30:14.746699] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:31:19.712 [2024-10-01 20:30:14.746745] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:19.712 [2024-10-01 20:30:14.749691] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:19.712 [2024-10-01 20:30:14.749757] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:31:19.712 BaseBdev1 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:19.712 BaseBdev2_malloc 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:19.712 [2024-10-01 20:30:14.812313] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:31:19.712 [2024-10-01 20:30:14.812427] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:19.712 [2024-10-01 20:30:14.812523] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:31:19.712 [2024-10-01 20:30:14.812556] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:19.712 [2024-10-01 20:30:14.816334] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:19.712 [2024-10-01 20:30:14.816389] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:31:19.712 BaseBdev2 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:19.712 BaseBdev3_malloc 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.712 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:19.712 [2024-10-01 20:30:14.877154] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:31:19.712 [2024-10-01 20:30:14.877257] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:19.713 [2024-10-01 20:30:14.877293] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:31:19.713 [2024-10-01 20:30:14.877314] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:19.713 [2024-10-01 20:30:14.880615] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:19.713 [2024-10-01 20:30:14.880701] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:31:19.713 BaseBdev3 00:31:19.713 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.713 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:31:19.713 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:31:19.713 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.713 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:19.713 BaseBdev4_malloc 00:31:19.713 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.713 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:31:19.713 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.713 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:19.713 [2024-10-01 20:30:14.942302] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:31:19.713 [2024-10-01 20:30:14.942533] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:19.713 [2024-10-01 20:30:14.942575] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:31:19.713 [2024-10-01 20:30:14.942603] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:19.713 [2024-10-01 20:30:14.945975] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:19.713 [2024-10-01 20:30:14.946050] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:31:19.713 BaseBdev4 00:31:19.713 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.713 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:31:19.713 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.713 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:19.971 spare_malloc 00:31:19.971 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.971 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:31:19.971 20:30:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:19.971 spare_delay 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:19.971 [2024-10-01 20:30:15.013453] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:31:19.971 [2024-10-01 20:30:15.013526] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:19.971 [2024-10-01 20:30:15.013558] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:31:19.971 [2024-10-01 20:30:15.013578] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:19.971 [2024-10-01 20:30:15.016574] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:19.971 [2024-10-01 20:30:15.016640] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:31:19.971 spare 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:19.971 [2024-10-01 20:30:15.021557] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:31:19.971 [2024-10-01 20:30:15.024070] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:31:19.971 [2024-10-01 20:30:15.024176] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:31:19.971 [2024-10-01 20:30:15.024279] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:31:19.971 [2024-10-01 20:30:15.024559] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:31:19.971 [2024-10-01 20:30:15.024583] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:31:19.971 [2024-10-01 20:30:15.024933] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:31:19.971 [2024-10-01 20:30:15.032265] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:31:19.971 [2024-10-01 20:30:15.032313] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:31:19.971 [2024-10-01 20:30:15.032626] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.971 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:19.972 "name": "raid_bdev1", 00:31:19.972 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:19.972 "strip_size_kb": 64, 00:31:19.972 "state": "online", 00:31:19.972 "raid_level": "raid5f", 00:31:19.972 "superblock": true, 00:31:19.972 "num_base_bdevs": 4, 00:31:19.972 "num_base_bdevs_discovered": 4, 00:31:19.972 "num_base_bdevs_operational": 4, 00:31:19.972 "base_bdevs_list": [ 00:31:19.972 { 00:31:19.972 "name": "BaseBdev1", 00:31:19.972 "uuid": "8c7c8eee-1595-57e6-956a-e3d4e71c0f2b", 00:31:19.972 "is_configured": true, 00:31:19.972 "data_offset": 2048, 00:31:19.972 "data_size": 63488 00:31:19.972 }, 00:31:19.972 { 00:31:19.972 "name": "BaseBdev2", 00:31:19.972 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:19.972 "is_configured": true, 00:31:19.972 "data_offset": 2048, 00:31:19.972 "data_size": 63488 00:31:19.972 }, 00:31:19.972 { 00:31:19.972 "name": "BaseBdev3", 00:31:19.972 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:19.972 "is_configured": true, 00:31:19.972 "data_offset": 2048, 00:31:19.972 "data_size": 63488 00:31:19.972 }, 00:31:19.972 { 00:31:19.972 "name": "BaseBdev4", 00:31:19.972 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:19.972 "is_configured": true, 00:31:19.972 "data_offset": 2048, 00:31:19.972 "data_size": 63488 00:31:19.972 } 00:31:19.972 ] 00:31:19.972 }' 00:31:19.972 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:19.972 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:20.537 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:20.538 [2024-10-01 20:30:15.584766] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=190464 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:31:20.538 20:30:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:31:20.796 [2024-10-01 20:30:15.988955] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:31:20.796 /dev/nbd0 00:31:20.796 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:31:20.796 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:31:20.796 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:31:20.796 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:31:20.796 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:31:20.796 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:31:20.796 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:31:20.796 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:31:20.796 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:31:20.796 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:31:20.796 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:20.796 1+0 records in 00:31:20.796 1+0 records out 00:31:20.796 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000331776 s, 12.3 MB/s 00:31:20.796 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:21.055 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:31:21.055 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:21.056 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:31:21.056 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:31:21.056 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:21.056 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:31:21.056 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:31:21.056 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:31:21.056 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 192 00:31:21.056 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=496 oflag=direct 00:31:21.623 496+0 records in 00:31:21.623 496+0 records out 00:31:21.623 97517568 bytes (98 MB, 93 MiB) copied, 0.680945 s, 143 MB/s 00:31:21.623 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:31:21.623 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:31:21.623 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:31:21.623 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:31:21.623 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:31:21.623 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:21.623 20:30:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:31:21.882 [2024-10-01 20:30:17.022843] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:21.882 [2024-10-01 20:30:17.048838] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:21.882 "name": "raid_bdev1", 00:31:21.882 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:21.882 "strip_size_kb": 64, 00:31:21.882 "state": "online", 00:31:21.882 "raid_level": "raid5f", 00:31:21.882 "superblock": true, 00:31:21.882 "num_base_bdevs": 4, 00:31:21.882 "num_base_bdevs_discovered": 3, 00:31:21.882 "num_base_bdevs_operational": 3, 00:31:21.882 "base_bdevs_list": [ 00:31:21.882 { 00:31:21.882 "name": null, 00:31:21.882 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:21.882 "is_configured": false, 00:31:21.882 "data_offset": 0, 00:31:21.882 "data_size": 63488 00:31:21.882 }, 00:31:21.882 { 00:31:21.882 "name": "BaseBdev2", 00:31:21.882 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:21.882 "is_configured": true, 00:31:21.882 "data_offset": 2048, 00:31:21.882 "data_size": 63488 00:31:21.882 }, 00:31:21.882 { 00:31:21.882 "name": "BaseBdev3", 00:31:21.882 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:21.882 "is_configured": true, 00:31:21.882 "data_offset": 2048, 00:31:21.882 "data_size": 63488 00:31:21.882 }, 00:31:21.882 { 00:31:21.882 "name": "BaseBdev4", 00:31:21.882 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:21.882 "is_configured": true, 00:31:21.882 "data_offset": 2048, 00:31:21.882 "data_size": 63488 00:31:21.882 } 00:31:21.882 ] 00:31:21.882 }' 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:21.882 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:22.448 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:31:22.448 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:22.448 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:22.448 [2024-10-01 20:30:17.629452] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:22.448 [2024-10-01 20:30:17.645965] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002aa50 00:31:22.448 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:22.448 20:30:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:31:22.448 [2024-10-01 20:30:17.656266] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:31:23.825 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:23.826 "name": "raid_bdev1", 00:31:23.826 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:23.826 "strip_size_kb": 64, 00:31:23.826 "state": "online", 00:31:23.826 "raid_level": "raid5f", 00:31:23.826 "superblock": true, 00:31:23.826 "num_base_bdevs": 4, 00:31:23.826 "num_base_bdevs_discovered": 4, 00:31:23.826 "num_base_bdevs_operational": 4, 00:31:23.826 "process": { 00:31:23.826 "type": "rebuild", 00:31:23.826 "target": "spare", 00:31:23.826 "progress": { 00:31:23.826 "blocks": 17280, 00:31:23.826 "percent": 9 00:31:23.826 } 00:31:23.826 }, 00:31:23.826 "base_bdevs_list": [ 00:31:23.826 { 00:31:23.826 "name": "spare", 00:31:23.826 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:23.826 "is_configured": true, 00:31:23.826 "data_offset": 2048, 00:31:23.826 "data_size": 63488 00:31:23.826 }, 00:31:23.826 { 00:31:23.826 "name": "BaseBdev2", 00:31:23.826 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:23.826 "is_configured": true, 00:31:23.826 "data_offset": 2048, 00:31:23.826 "data_size": 63488 00:31:23.826 }, 00:31:23.826 { 00:31:23.826 "name": "BaseBdev3", 00:31:23.826 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:23.826 "is_configured": true, 00:31:23.826 "data_offset": 2048, 00:31:23.826 "data_size": 63488 00:31:23.826 }, 00:31:23.826 { 00:31:23.826 "name": "BaseBdev4", 00:31:23.826 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:23.826 "is_configured": true, 00:31:23.826 "data_offset": 2048, 00:31:23.826 "data_size": 63488 00:31:23.826 } 00:31:23.826 ] 00:31:23.826 }' 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:23.826 [2024-10-01 20:30:18.822253] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:23.826 [2024-10-01 20:30:18.869953] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:31:23.826 [2024-10-01 20:30:18.870201] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:23.826 [2024-10-01 20:30:18.870234] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:23.826 [2024-10-01 20:30:18.870255] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:23.826 "name": "raid_bdev1", 00:31:23.826 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:23.826 "strip_size_kb": 64, 00:31:23.826 "state": "online", 00:31:23.826 "raid_level": "raid5f", 00:31:23.826 "superblock": true, 00:31:23.826 "num_base_bdevs": 4, 00:31:23.826 "num_base_bdevs_discovered": 3, 00:31:23.826 "num_base_bdevs_operational": 3, 00:31:23.826 "base_bdevs_list": [ 00:31:23.826 { 00:31:23.826 "name": null, 00:31:23.826 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:23.826 "is_configured": false, 00:31:23.826 "data_offset": 0, 00:31:23.826 "data_size": 63488 00:31:23.826 }, 00:31:23.826 { 00:31:23.826 "name": "BaseBdev2", 00:31:23.826 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:23.826 "is_configured": true, 00:31:23.826 "data_offset": 2048, 00:31:23.826 "data_size": 63488 00:31:23.826 }, 00:31:23.826 { 00:31:23.826 "name": "BaseBdev3", 00:31:23.826 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:23.826 "is_configured": true, 00:31:23.826 "data_offset": 2048, 00:31:23.826 "data_size": 63488 00:31:23.826 }, 00:31:23.826 { 00:31:23.826 "name": "BaseBdev4", 00:31:23.826 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:23.826 "is_configured": true, 00:31:23.826 "data_offset": 2048, 00:31:23.826 "data_size": 63488 00:31:23.826 } 00:31:23.826 ] 00:31:23.826 }' 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:23.826 20:30:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:24.396 "name": "raid_bdev1", 00:31:24.396 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:24.396 "strip_size_kb": 64, 00:31:24.396 "state": "online", 00:31:24.396 "raid_level": "raid5f", 00:31:24.396 "superblock": true, 00:31:24.396 "num_base_bdevs": 4, 00:31:24.396 "num_base_bdevs_discovered": 3, 00:31:24.396 "num_base_bdevs_operational": 3, 00:31:24.396 "base_bdevs_list": [ 00:31:24.396 { 00:31:24.396 "name": null, 00:31:24.396 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:24.396 "is_configured": false, 00:31:24.396 "data_offset": 0, 00:31:24.396 "data_size": 63488 00:31:24.396 }, 00:31:24.396 { 00:31:24.396 "name": "BaseBdev2", 00:31:24.396 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:24.396 "is_configured": true, 00:31:24.396 "data_offset": 2048, 00:31:24.396 "data_size": 63488 00:31:24.396 }, 00:31:24.396 { 00:31:24.396 "name": "BaseBdev3", 00:31:24.396 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:24.396 "is_configured": true, 00:31:24.396 "data_offset": 2048, 00:31:24.396 "data_size": 63488 00:31:24.396 }, 00:31:24.396 { 00:31:24.396 "name": "BaseBdev4", 00:31:24.396 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:24.396 "is_configured": true, 00:31:24.396 "data_offset": 2048, 00:31:24.396 "data_size": 63488 00:31:24.396 } 00:31:24.396 ] 00:31:24.396 }' 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:24.396 [2024-10-01 20:30:19.605222] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:24.396 [2024-10-01 20:30:19.621185] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00002ab20 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:24.396 20:30:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:31:24.396 [2024-10-01 20:30:19.632187] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:31:25.771 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:25.771 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:25.771 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:25.771 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:25.771 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:25.771 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:25.771 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:25.771 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:25.771 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:25.771 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:25.771 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:25.771 "name": "raid_bdev1", 00:31:25.771 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:25.771 "strip_size_kb": 64, 00:31:25.771 "state": "online", 00:31:25.771 "raid_level": "raid5f", 00:31:25.771 "superblock": true, 00:31:25.771 "num_base_bdevs": 4, 00:31:25.771 "num_base_bdevs_discovered": 4, 00:31:25.771 "num_base_bdevs_operational": 4, 00:31:25.771 "process": { 00:31:25.771 "type": "rebuild", 00:31:25.771 "target": "spare", 00:31:25.771 "progress": { 00:31:25.771 "blocks": 17280, 00:31:25.771 "percent": 9 00:31:25.771 } 00:31:25.771 }, 00:31:25.771 "base_bdevs_list": [ 00:31:25.771 { 00:31:25.771 "name": "spare", 00:31:25.771 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:25.771 "is_configured": true, 00:31:25.771 "data_offset": 2048, 00:31:25.771 "data_size": 63488 00:31:25.771 }, 00:31:25.771 { 00:31:25.771 "name": "BaseBdev2", 00:31:25.771 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:25.771 "is_configured": true, 00:31:25.771 "data_offset": 2048, 00:31:25.771 "data_size": 63488 00:31:25.771 }, 00:31:25.771 { 00:31:25.771 "name": "BaseBdev3", 00:31:25.771 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:25.771 "is_configured": true, 00:31:25.771 "data_offset": 2048, 00:31:25.771 "data_size": 63488 00:31:25.771 }, 00:31:25.771 { 00:31:25.771 "name": "BaseBdev4", 00:31:25.772 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:25.772 "is_configured": true, 00:31:25.772 "data_offset": 2048, 00:31:25.772 "data_size": 63488 00:31:25.772 } 00:31:25.772 ] 00:31:25.772 }' 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:31:25.772 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=759 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:25.772 "name": "raid_bdev1", 00:31:25.772 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:25.772 "strip_size_kb": 64, 00:31:25.772 "state": "online", 00:31:25.772 "raid_level": "raid5f", 00:31:25.772 "superblock": true, 00:31:25.772 "num_base_bdevs": 4, 00:31:25.772 "num_base_bdevs_discovered": 4, 00:31:25.772 "num_base_bdevs_operational": 4, 00:31:25.772 "process": { 00:31:25.772 "type": "rebuild", 00:31:25.772 "target": "spare", 00:31:25.772 "progress": { 00:31:25.772 "blocks": 21120, 00:31:25.772 "percent": 11 00:31:25.772 } 00:31:25.772 }, 00:31:25.772 "base_bdevs_list": [ 00:31:25.772 { 00:31:25.772 "name": "spare", 00:31:25.772 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:25.772 "is_configured": true, 00:31:25.772 "data_offset": 2048, 00:31:25.772 "data_size": 63488 00:31:25.772 }, 00:31:25.772 { 00:31:25.772 "name": "BaseBdev2", 00:31:25.772 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:25.772 "is_configured": true, 00:31:25.772 "data_offset": 2048, 00:31:25.772 "data_size": 63488 00:31:25.772 }, 00:31:25.772 { 00:31:25.772 "name": "BaseBdev3", 00:31:25.772 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:25.772 "is_configured": true, 00:31:25.772 "data_offset": 2048, 00:31:25.772 "data_size": 63488 00:31:25.772 }, 00:31:25.772 { 00:31:25.772 "name": "BaseBdev4", 00:31:25.772 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:25.772 "is_configured": true, 00:31:25.772 "data_offset": 2048, 00:31:25.772 "data_size": 63488 00:31:25.772 } 00:31:25.772 ] 00:31:25.772 }' 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:25.772 20:30:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:27.145 20:30:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:27.145 20:30:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:27.145 20:30:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:27.145 20:30:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:27.145 20:30:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:27.145 20:30:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:27.145 20:30:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:27.145 20:30:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:27.145 20:30:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:27.145 20:30:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:27.145 20:30:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:27.145 20:30:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:27.145 "name": "raid_bdev1", 00:31:27.145 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:27.145 "strip_size_kb": 64, 00:31:27.145 "state": "online", 00:31:27.145 "raid_level": "raid5f", 00:31:27.145 "superblock": true, 00:31:27.145 "num_base_bdevs": 4, 00:31:27.145 "num_base_bdevs_discovered": 4, 00:31:27.145 "num_base_bdevs_operational": 4, 00:31:27.145 "process": { 00:31:27.145 "type": "rebuild", 00:31:27.145 "target": "spare", 00:31:27.145 "progress": { 00:31:27.145 "blocks": 44160, 00:31:27.145 "percent": 23 00:31:27.145 } 00:31:27.145 }, 00:31:27.145 "base_bdevs_list": [ 00:31:27.145 { 00:31:27.145 "name": "spare", 00:31:27.146 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:27.146 "is_configured": true, 00:31:27.146 "data_offset": 2048, 00:31:27.146 "data_size": 63488 00:31:27.146 }, 00:31:27.146 { 00:31:27.146 "name": "BaseBdev2", 00:31:27.146 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:27.146 "is_configured": true, 00:31:27.146 "data_offset": 2048, 00:31:27.146 "data_size": 63488 00:31:27.146 }, 00:31:27.146 { 00:31:27.146 "name": "BaseBdev3", 00:31:27.146 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:27.146 "is_configured": true, 00:31:27.146 "data_offset": 2048, 00:31:27.146 "data_size": 63488 00:31:27.146 }, 00:31:27.146 { 00:31:27.146 "name": "BaseBdev4", 00:31:27.146 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:27.146 "is_configured": true, 00:31:27.146 "data_offset": 2048, 00:31:27.146 "data_size": 63488 00:31:27.146 } 00:31:27.146 ] 00:31:27.146 }' 00:31:27.146 20:30:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:27.146 20:30:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:27.146 20:30:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:27.146 20:30:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:27.146 20:30:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:28.079 "name": "raid_bdev1", 00:31:28.079 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:28.079 "strip_size_kb": 64, 00:31:28.079 "state": "online", 00:31:28.079 "raid_level": "raid5f", 00:31:28.079 "superblock": true, 00:31:28.079 "num_base_bdevs": 4, 00:31:28.079 "num_base_bdevs_discovered": 4, 00:31:28.079 "num_base_bdevs_operational": 4, 00:31:28.079 "process": { 00:31:28.079 "type": "rebuild", 00:31:28.079 "target": "spare", 00:31:28.079 "progress": { 00:31:28.079 "blocks": 65280, 00:31:28.079 "percent": 34 00:31:28.079 } 00:31:28.079 }, 00:31:28.079 "base_bdevs_list": [ 00:31:28.079 { 00:31:28.079 "name": "spare", 00:31:28.079 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:28.079 "is_configured": true, 00:31:28.079 "data_offset": 2048, 00:31:28.079 "data_size": 63488 00:31:28.079 }, 00:31:28.079 { 00:31:28.079 "name": "BaseBdev2", 00:31:28.079 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:28.079 "is_configured": true, 00:31:28.079 "data_offset": 2048, 00:31:28.079 "data_size": 63488 00:31:28.079 }, 00:31:28.079 { 00:31:28.079 "name": "BaseBdev3", 00:31:28.079 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:28.079 "is_configured": true, 00:31:28.079 "data_offset": 2048, 00:31:28.079 "data_size": 63488 00:31:28.079 }, 00:31:28.079 { 00:31:28.079 "name": "BaseBdev4", 00:31:28.079 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:28.079 "is_configured": true, 00:31:28.079 "data_offset": 2048, 00:31:28.079 "data_size": 63488 00:31:28.079 } 00:31:28.079 ] 00:31:28.079 }' 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:28.079 20:30:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:29.452 "name": "raid_bdev1", 00:31:29.452 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:29.452 "strip_size_kb": 64, 00:31:29.452 "state": "online", 00:31:29.452 "raid_level": "raid5f", 00:31:29.452 "superblock": true, 00:31:29.452 "num_base_bdevs": 4, 00:31:29.452 "num_base_bdevs_discovered": 4, 00:31:29.452 "num_base_bdevs_operational": 4, 00:31:29.452 "process": { 00:31:29.452 "type": "rebuild", 00:31:29.452 "target": "spare", 00:31:29.452 "progress": { 00:31:29.452 "blocks": 88320, 00:31:29.452 "percent": 46 00:31:29.452 } 00:31:29.452 }, 00:31:29.452 "base_bdevs_list": [ 00:31:29.452 { 00:31:29.452 "name": "spare", 00:31:29.452 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:29.452 "is_configured": true, 00:31:29.452 "data_offset": 2048, 00:31:29.452 "data_size": 63488 00:31:29.452 }, 00:31:29.452 { 00:31:29.452 "name": "BaseBdev2", 00:31:29.452 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:29.452 "is_configured": true, 00:31:29.452 "data_offset": 2048, 00:31:29.452 "data_size": 63488 00:31:29.452 }, 00:31:29.452 { 00:31:29.452 "name": "BaseBdev3", 00:31:29.452 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:29.452 "is_configured": true, 00:31:29.452 "data_offset": 2048, 00:31:29.452 "data_size": 63488 00:31:29.452 }, 00:31:29.452 { 00:31:29.452 "name": "BaseBdev4", 00:31:29.452 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:29.452 "is_configured": true, 00:31:29.452 "data_offset": 2048, 00:31:29.452 "data_size": 63488 00:31:29.452 } 00:31:29.452 ] 00:31:29.452 }' 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:29.452 20:30:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:30.387 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:30.387 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:30.387 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:30.387 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:30.387 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:30.387 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:30.387 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:30.387 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:30.387 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:30.387 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:30.387 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:30.387 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:30.387 "name": "raid_bdev1", 00:31:30.387 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:30.387 "strip_size_kb": 64, 00:31:30.387 "state": "online", 00:31:30.387 "raid_level": "raid5f", 00:31:30.387 "superblock": true, 00:31:30.387 "num_base_bdevs": 4, 00:31:30.387 "num_base_bdevs_discovered": 4, 00:31:30.387 "num_base_bdevs_operational": 4, 00:31:30.387 "process": { 00:31:30.387 "type": "rebuild", 00:31:30.387 "target": "spare", 00:31:30.387 "progress": { 00:31:30.387 "blocks": 111360, 00:31:30.387 "percent": 58 00:31:30.387 } 00:31:30.387 }, 00:31:30.387 "base_bdevs_list": [ 00:31:30.387 { 00:31:30.387 "name": "spare", 00:31:30.387 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:30.387 "is_configured": true, 00:31:30.387 "data_offset": 2048, 00:31:30.387 "data_size": 63488 00:31:30.387 }, 00:31:30.387 { 00:31:30.387 "name": "BaseBdev2", 00:31:30.387 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:30.387 "is_configured": true, 00:31:30.387 "data_offset": 2048, 00:31:30.387 "data_size": 63488 00:31:30.387 }, 00:31:30.387 { 00:31:30.387 "name": "BaseBdev3", 00:31:30.387 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:30.387 "is_configured": true, 00:31:30.387 "data_offset": 2048, 00:31:30.387 "data_size": 63488 00:31:30.387 }, 00:31:30.387 { 00:31:30.387 "name": "BaseBdev4", 00:31:30.387 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:30.387 "is_configured": true, 00:31:30.387 "data_offset": 2048, 00:31:30.387 "data_size": 63488 00:31:30.387 } 00:31:30.387 ] 00:31:30.387 }' 00:31:30.387 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:30.387 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:30.387 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:30.645 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:30.645 20:30:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:31.583 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:31.583 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:31.583 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:31.583 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:31.583 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:31.583 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:31.583 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:31.583 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:31.583 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:31.583 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:31.583 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:31.583 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:31.583 "name": "raid_bdev1", 00:31:31.583 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:31.583 "strip_size_kb": 64, 00:31:31.583 "state": "online", 00:31:31.583 "raid_level": "raid5f", 00:31:31.583 "superblock": true, 00:31:31.583 "num_base_bdevs": 4, 00:31:31.583 "num_base_bdevs_discovered": 4, 00:31:31.583 "num_base_bdevs_operational": 4, 00:31:31.583 "process": { 00:31:31.583 "type": "rebuild", 00:31:31.583 "target": "spare", 00:31:31.583 "progress": { 00:31:31.583 "blocks": 132480, 00:31:31.583 "percent": 69 00:31:31.583 } 00:31:31.583 }, 00:31:31.583 "base_bdevs_list": [ 00:31:31.583 { 00:31:31.583 "name": "spare", 00:31:31.583 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:31.583 "is_configured": true, 00:31:31.583 "data_offset": 2048, 00:31:31.583 "data_size": 63488 00:31:31.583 }, 00:31:31.583 { 00:31:31.583 "name": "BaseBdev2", 00:31:31.584 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:31.584 "is_configured": true, 00:31:31.584 "data_offset": 2048, 00:31:31.584 "data_size": 63488 00:31:31.584 }, 00:31:31.584 { 00:31:31.584 "name": "BaseBdev3", 00:31:31.584 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:31.584 "is_configured": true, 00:31:31.584 "data_offset": 2048, 00:31:31.584 "data_size": 63488 00:31:31.584 }, 00:31:31.584 { 00:31:31.584 "name": "BaseBdev4", 00:31:31.584 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:31.584 "is_configured": true, 00:31:31.584 "data_offset": 2048, 00:31:31.584 "data_size": 63488 00:31:31.584 } 00:31:31.584 ] 00:31:31.584 }' 00:31:31.584 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:31.584 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:31.584 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:31.906 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:31.906 20:30:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:32.840 20:30:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:32.840 20:30:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:32.840 20:30:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:32.840 20:30:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:32.840 20:30:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:32.840 20:30:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:32.840 20:30:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:32.840 20:30:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:32.840 20:30:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:32.840 20:30:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:32.840 20:30:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:32.840 20:30:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:32.840 "name": "raid_bdev1", 00:31:32.840 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:32.840 "strip_size_kb": 64, 00:31:32.840 "state": "online", 00:31:32.840 "raid_level": "raid5f", 00:31:32.840 "superblock": true, 00:31:32.840 "num_base_bdevs": 4, 00:31:32.840 "num_base_bdevs_discovered": 4, 00:31:32.840 "num_base_bdevs_operational": 4, 00:31:32.840 "process": { 00:31:32.840 "type": "rebuild", 00:31:32.840 "target": "spare", 00:31:32.840 "progress": { 00:31:32.840 "blocks": 155520, 00:31:32.840 "percent": 81 00:31:32.840 } 00:31:32.840 }, 00:31:32.840 "base_bdevs_list": [ 00:31:32.840 { 00:31:32.840 "name": "spare", 00:31:32.840 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:32.840 "is_configured": true, 00:31:32.840 "data_offset": 2048, 00:31:32.840 "data_size": 63488 00:31:32.840 }, 00:31:32.840 { 00:31:32.840 "name": "BaseBdev2", 00:31:32.840 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:32.840 "is_configured": true, 00:31:32.840 "data_offset": 2048, 00:31:32.840 "data_size": 63488 00:31:32.840 }, 00:31:32.840 { 00:31:32.840 "name": "BaseBdev3", 00:31:32.840 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:32.840 "is_configured": true, 00:31:32.840 "data_offset": 2048, 00:31:32.840 "data_size": 63488 00:31:32.840 }, 00:31:32.840 { 00:31:32.840 "name": "BaseBdev4", 00:31:32.840 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:32.840 "is_configured": true, 00:31:32.840 "data_offset": 2048, 00:31:32.840 "data_size": 63488 00:31:32.841 } 00:31:32.841 ] 00:31:32.841 }' 00:31:32.841 20:30:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:32.841 20:30:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:32.841 20:30:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:32.841 20:30:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:32.841 20:30:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:34.214 "name": "raid_bdev1", 00:31:34.214 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:34.214 "strip_size_kb": 64, 00:31:34.214 "state": "online", 00:31:34.214 "raid_level": "raid5f", 00:31:34.214 "superblock": true, 00:31:34.214 "num_base_bdevs": 4, 00:31:34.214 "num_base_bdevs_discovered": 4, 00:31:34.214 "num_base_bdevs_operational": 4, 00:31:34.214 "process": { 00:31:34.214 "type": "rebuild", 00:31:34.214 "target": "spare", 00:31:34.214 "progress": { 00:31:34.214 "blocks": 176640, 00:31:34.214 "percent": 92 00:31:34.214 } 00:31:34.214 }, 00:31:34.214 "base_bdevs_list": [ 00:31:34.214 { 00:31:34.214 "name": "spare", 00:31:34.214 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:34.214 "is_configured": true, 00:31:34.214 "data_offset": 2048, 00:31:34.214 "data_size": 63488 00:31:34.214 }, 00:31:34.214 { 00:31:34.214 "name": "BaseBdev2", 00:31:34.214 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:34.214 "is_configured": true, 00:31:34.214 "data_offset": 2048, 00:31:34.214 "data_size": 63488 00:31:34.214 }, 00:31:34.214 { 00:31:34.214 "name": "BaseBdev3", 00:31:34.214 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:34.214 "is_configured": true, 00:31:34.214 "data_offset": 2048, 00:31:34.214 "data_size": 63488 00:31:34.214 }, 00:31:34.214 { 00:31:34.214 "name": "BaseBdev4", 00:31:34.214 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:34.214 "is_configured": true, 00:31:34.214 "data_offset": 2048, 00:31:34.214 "data_size": 63488 00:31:34.214 } 00:31:34.214 ] 00:31:34.214 }' 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:34.214 20:30:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:31:34.781 [2024-10-01 20:30:29.749213] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:31:34.781 [2024-10-01 20:30:29.749399] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:31:34.781 [2024-10-01 20:30:29.749626] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:35.039 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:31:35.039 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:35.039 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:35.039 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:35.039 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:35.039 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:35.039 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:35.039 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:35.039 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:35.039 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:35.039 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:35.039 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:35.039 "name": "raid_bdev1", 00:31:35.039 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:35.039 "strip_size_kb": 64, 00:31:35.039 "state": "online", 00:31:35.039 "raid_level": "raid5f", 00:31:35.039 "superblock": true, 00:31:35.039 "num_base_bdevs": 4, 00:31:35.039 "num_base_bdevs_discovered": 4, 00:31:35.039 "num_base_bdevs_operational": 4, 00:31:35.039 "base_bdevs_list": [ 00:31:35.039 { 00:31:35.039 "name": "spare", 00:31:35.039 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:35.039 "is_configured": true, 00:31:35.039 "data_offset": 2048, 00:31:35.039 "data_size": 63488 00:31:35.039 }, 00:31:35.039 { 00:31:35.039 "name": "BaseBdev2", 00:31:35.039 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:35.039 "is_configured": true, 00:31:35.039 "data_offset": 2048, 00:31:35.039 "data_size": 63488 00:31:35.039 }, 00:31:35.039 { 00:31:35.039 "name": "BaseBdev3", 00:31:35.039 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:35.039 "is_configured": true, 00:31:35.039 "data_offset": 2048, 00:31:35.039 "data_size": 63488 00:31:35.039 }, 00:31:35.039 { 00:31:35.039 "name": "BaseBdev4", 00:31:35.039 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:35.039 "is_configured": true, 00:31:35.039 "data_offset": 2048, 00:31:35.039 "data_size": 63488 00:31:35.039 } 00:31:35.039 ] 00:31:35.039 }' 00:31:35.039 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:35.298 "name": "raid_bdev1", 00:31:35.298 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:35.298 "strip_size_kb": 64, 00:31:35.298 "state": "online", 00:31:35.298 "raid_level": "raid5f", 00:31:35.298 "superblock": true, 00:31:35.298 "num_base_bdevs": 4, 00:31:35.298 "num_base_bdevs_discovered": 4, 00:31:35.298 "num_base_bdevs_operational": 4, 00:31:35.298 "base_bdevs_list": [ 00:31:35.298 { 00:31:35.298 "name": "spare", 00:31:35.298 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:35.298 "is_configured": true, 00:31:35.298 "data_offset": 2048, 00:31:35.298 "data_size": 63488 00:31:35.298 }, 00:31:35.298 { 00:31:35.298 "name": "BaseBdev2", 00:31:35.298 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:35.298 "is_configured": true, 00:31:35.298 "data_offset": 2048, 00:31:35.298 "data_size": 63488 00:31:35.298 }, 00:31:35.298 { 00:31:35.298 "name": "BaseBdev3", 00:31:35.298 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:35.298 "is_configured": true, 00:31:35.298 "data_offset": 2048, 00:31:35.298 "data_size": 63488 00:31:35.298 }, 00:31:35.298 { 00:31:35.298 "name": "BaseBdev4", 00:31:35.298 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:35.298 "is_configured": true, 00:31:35.298 "data_offset": 2048, 00:31:35.298 "data_size": 63488 00:31:35.298 } 00:31:35.298 ] 00:31:35.298 }' 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:35.298 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:35.556 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:35.556 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:35.556 "name": "raid_bdev1", 00:31:35.556 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:35.556 "strip_size_kb": 64, 00:31:35.556 "state": "online", 00:31:35.556 "raid_level": "raid5f", 00:31:35.556 "superblock": true, 00:31:35.556 "num_base_bdevs": 4, 00:31:35.556 "num_base_bdevs_discovered": 4, 00:31:35.556 "num_base_bdevs_operational": 4, 00:31:35.556 "base_bdevs_list": [ 00:31:35.556 { 00:31:35.556 "name": "spare", 00:31:35.556 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:35.556 "is_configured": true, 00:31:35.556 "data_offset": 2048, 00:31:35.556 "data_size": 63488 00:31:35.556 }, 00:31:35.556 { 00:31:35.556 "name": "BaseBdev2", 00:31:35.556 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:35.556 "is_configured": true, 00:31:35.556 "data_offset": 2048, 00:31:35.556 "data_size": 63488 00:31:35.556 }, 00:31:35.556 { 00:31:35.556 "name": "BaseBdev3", 00:31:35.556 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:35.556 "is_configured": true, 00:31:35.556 "data_offset": 2048, 00:31:35.556 "data_size": 63488 00:31:35.556 }, 00:31:35.556 { 00:31:35.556 "name": "BaseBdev4", 00:31:35.556 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:35.556 "is_configured": true, 00:31:35.556 "data_offset": 2048, 00:31:35.556 "data_size": 63488 00:31:35.556 } 00:31:35.556 ] 00:31:35.556 }' 00:31:35.556 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:35.556 20:30:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:36.122 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:31:36.122 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:36.122 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:36.122 [2024-10-01 20:30:31.100262] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:31:36.122 [2024-10-01 20:30:31.100331] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:31:36.122 [2024-10-01 20:30:31.100528] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:31:36.122 [2024-10-01 20:30:31.100714] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:31:36.123 [2024-10-01 20:30:31.100763] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:31:36.123 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:31:36.381 /dev/nbd0 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:36.381 1+0 records in 00:31:36.381 1+0 records out 00:31:36.381 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000255312 s, 16.0 MB/s 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:31:36.381 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:31:36.640 /dev/nbd1 00:31:36.640 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:31:36.640 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:31:36.640 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:31:36.640 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:31:36.640 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:31:36.640 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:31:36.640 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:31:36.640 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:31:36.640 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:31:36.640 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:31:36.640 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:31:36.640 1+0 records in 00:31:36.640 1+0 records out 00:31:36.640 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000425062 s, 9.6 MB/s 00:31:36.640 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:36.898 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:31:36.898 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:31:36.898 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:31:36.898 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:31:36.898 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:31:36.898 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:31:36.898 20:30:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:31:36.898 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:31:36.898 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:31:36.898 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:31:36.898 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:31:36.898 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:31:36.898 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:36.898 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:31:37.464 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:31:37.464 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:31:37.464 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:31:37.464 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:37.464 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:37.464 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:31:37.464 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:31:37.464 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:31:37.464 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:31:37.464 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:37.723 [2024-10-01 20:30:32.777010] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:31:37.723 [2024-10-01 20:30:32.777121] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:37.723 [2024-10-01 20:30:32.777195] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:31:37.723 [2024-10-01 20:30:32.777228] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:37.723 [2024-10-01 20:30:32.781078] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:37.723 [2024-10-01 20:30:32.781132] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:31:37.723 [2024-10-01 20:30:32.781414] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:31:37.723 [2024-10-01 20:30:32.781515] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:37.723 [2024-10-01 20:30:32.781852] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:31:37.723 [2024-10-01 20:30:32.782099] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:31:37.723 [2024-10-01 20:30:32.782282] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:31:37.723 spare 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:37.723 [2024-10-01 20:30:32.882503] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:31:37.723 [2024-10-01 20:30:32.882639] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:31:37.723 [2024-10-01 20:30:32.883410] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000491d0 00:31:37.723 [2024-10-01 20:30:32.891273] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:31:37.723 [2024-10-01 20:30:32.891326] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:31:37.723 [2024-10-01 20:30:32.891732] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:37.723 "name": "raid_bdev1", 00:31:37.723 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:37.723 "strip_size_kb": 64, 00:31:37.723 "state": "online", 00:31:37.723 "raid_level": "raid5f", 00:31:37.723 "superblock": true, 00:31:37.723 "num_base_bdevs": 4, 00:31:37.723 "num_base_bdevs_discovered": 4, 00:31:37.723 "num_base_bdevs_operational": 4, 00:31:37.723 "base_bdevs_list": [ 00:31:37.723 { 00:31:37.723 "name": "spare", 00:31:37.723 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:37.723 "is_configured": true, 00:31:37.723 "data_offset": 2048, 00:31:37.723 "data_size": 63488 00:31:37.723 }, 00:31:37.723 { 00:31:37.723 "name": "BaseBdev2", 00:31:37.723 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:37.723 "is_configured": true, 00:31:37.723 "data_offset": 2048, 00:31:37.723 "data_size": 63488 00:31:37.723 }, 00:31:37.723 { 00:31:37.723 "name": "BaseBdev3", 00:31:37.723 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:37.723 "is_configured": true, 00:31:37.723 "data_offset": 2048, 00:31:37.723 "data_size": 63488 00:31:37.723 }, 00:31:37.723 { 00:31:37.723 "name": "BaseBdev4", 00:31:37.723 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:37.723 "is_configured": true, 00:31:37.723 "data_offset": 2048, 00:31:37.723 "data_size": 63488 00:31:37.723 } 00:31:37.723 ] 00:31:37.723 }' 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:37.723 20:30:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:38.290 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:38.290 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:38.290 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:38.290 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:38.290 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:38.290 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:38.290 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:38.290 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:38.290 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:38.290 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:38.290 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:38.290 "name": "raid_bdev1", 00:31:38.290 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:38.290 "strip_size_kb": 64, 00:31:38.290 "state": "online", 00:31:38.290 "raid_level": "raid5f", 00:31:38.290 "superblock": true, 00:31:38.290 "num_base_bdevs": 4, 00:31:38.290 "num_base_bdevs_discovered": 4, 00:31:38.290 "num_base_bdevs_operational": 4, 00:31:38.290 "base_bdevs_list": [ 00:31:38.290 { 00:31:38.290 "name": "spare", 00:31:38.290 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:38.290 "is_configured": true, 00:31:38.290 "data_offset": 2048, 00:31:38.290 "data_size": 63488 00:31:38.290 }, 00:31:38.290 { 00:31:38.290 "name": "BaseBdev2", 00:31:38.290 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:38.290 "is_configured": true, 00:31:38.290 "data_offset": 2048, 00:31:38.290 "data_size": 63488 00:31:38.290 }, 00:31:38.290 { 00:31:38.290 "name": "BaseBdev3", 00:31:38.290 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:38.290 "is_configured": true, 00:31:38.290 "data_offset": 2048, 00:31:38.290 "data_size": 63488 00:31:38.290 }, 00:31:38.290 { 00:31:38.290 "name": "BaseBdev4", 00:31:38.290 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:38.290 "is_configured": true, 00:31:38.290 "data_offset": 2048, 00:31:38.290 "data_size": 63488 00:31:38.290 } 00:31:38.290 ] 00:31:38.290 }' 00:31:38.290 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:38.290 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:38.290 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:38.549 [2024-10-01 20:30:33.645433] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:38.549 "name": "raid_bdev1", 00:31:38.549 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:38.549 "strip_size_kb": 64, 00:31:38.549 "state": "online", 00:31:38.549 "raid_level": "raid5f", 00:31:38.549 "superblock": true, 00:31:38.549 "num_base_bdevs": 4, 00:31:38.549 "num_base_bdevs_discovered": 3, 00:31:38.549 "num_base_bdevs_operational": 3, 00:31:38.549 "base_bdevs_list": [ 00:31:38.549 { 00:31:38.549 "name": null, 00:31:38.549 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:38.549 "is_configured": false, 00:31:38.549 "data_offset": 0, 00:31:38.549 "data_size": 63488 00:31:38.549 }, 00:31:38.549 { 00:31:38.549 "name": "BaseBdev2", 00:31:38.549 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:38.549 "is_configured": true, 00:31:38.549 "data_offset": 2048, 00:31:38.549 "data_size": 63488 00:31:38.549 }, 00:31:38.549 { 00:31:38.549 "name": "BaseBdev3", 00:31:38.549 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:38.549 "is_configured": true, 00:31:38.549 "data_offset": 2048, 00:31:38.549 "data_size": 63488 00:31:38.549 }, 00:31:38.549 { 00:31:38.549 "name": "BaseBdev4", 00:31:38.549 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:38.549 "is_configured": true, 00:31:38.549 "data_offset": 2048, 00:31:38.549 "data_size": 63488 00:31:38.549 } 00:31:38.549 ] 00:31:38.549 }' 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:38.549 20:30:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:39.114 20:30:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:31:39.114 20:30:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:39.114 20:30:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:39.114 [2024-10-01 20:30:34.181618] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:39.114 [2024-10-01 20:30:34.181927] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:31:39.114 [2024-10-01 20:30:34.181976] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:31:39.114 [2024-10-01 20:30:34.182055] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:39.114 [2024-10-01 20:30:34.194792] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000492a0 00:31:39.114 20:30:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:39.114 20:30:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:31:39.114 [2024-10-01 20:30:34.204080] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:31:40.074 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:40.074 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:40.074 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:40.074 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:40.074 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:40.074 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:40.074 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:40.074 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:40.074 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:40.074 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:40.074 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:40.074 "name": "raid_bdev1", 00:31:40.074 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:40.074 "strip_size_kb": 64, 00:31:40.075 "state": "online", 00:31:40.075 "raid_level": "raid5f", 00:31:40.075 "superblock": true, 00:31:40.075 "num_base_bdevs": 4, 00:31:40.075 "num_base_bdevs_discovered": 4, 00:31:40.075 "num_base_bdevs_operational": 4, 00:31:40.075 "process": { 00:31:40.075 "type": "rebuild", 00:31:40.075 "target": "spare", 00:31:40.075 "progress": { 00:31:40.075 "blocks": 17280, 00:31:40.075 "percent": 9 00:31:40.075 } 00:31:40.075 }, 00:31:40.075 "base_bdevs_list": [ 00:31:40.075 { 00:31:40.075 "name": "spare", 00:31:40.075 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:40.075 "is_configured": true, 00:31:40.075 "data_offset": 2048, 00:31:40.075 "data_size": 63488 00:31:40.075 }, 00:31:40.075 { 00:31:40.075 "name": "BaseBdev2", 00:31:40.075 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:40.075 "is_configured": true, 00:31:40.075 "data_offset": 2048, 00:31:40.075 "data_size": 63488 00:31:40.075 }, 00:31:40.075 { 00:31:40.075 "name": "BaseBdev3", 00:31:40.075 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:40.075 "is_configured": true, 00:31:40.075 "data_offset": 2048, 00:31:40.075 "data_size": 63488 00:31:40.075 }, 00:31:40.075 { 00:31:40.075 "name": "BaseBdev4", 00:31:40.075 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:40.075 "is_configured": true, 00:31:40.075 "data_offset": 2048, 00:31:40.075 "data_size": 63488 00:31:40.075 } 00:31:40.075 ] 00:31:40.075 }' 00:31:40.075 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:40.075 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:40.075 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:40.334 [2024-10-01 20:30:35.358324] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:40.334 [2024-10-01 20:30:35.416789] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:31:40.334 [2024-10-01 20:30:35.416925] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:40.334 [2024-10-01 20:30:35.416972] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:40.334 [2024-10-01 20:30:35.417003] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:40.334 "name": "raid_bdev1", 00:31:40.334 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:40.334 "strip_size_kb": 64, 00:31:40.334 "state": "online", 00:31:40.334 "raid_level": "raid5f", 00:31:40.334 "superblock": true, 00:31:40.334 "num_base_bdevs": 4, 00:31:40.334 "num_base_bdevs_discovered": 3, 00:31:40.334 "num_base_bdevs_operational": 3, 00:31:40.334 "base_bdevs_list": [ 00:31:40.334 { 00:31:40.334 "name": null, 00:31:40.334 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:40.334 "is_configured": false, 00:31:40.334 "data_offset": 0, 00:31:40.334 "data_size": 63488 00:31:40.334 }, 00:31:40.334 { 00:31:40.334 "name": "BaseBdev2", 00:31:40.334 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:40.334 "is_configured": true, 00:31:40.334 "data_offset": 2048, 00:31:40.334 "data_size": 63488 00:31:40.334 }, 00:31:40.334 { 00:31:40.334 "name": "BaseBdev3", 00:31:40.334 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:40.334 "is_configured": true, 00:31:40.334 "data_offset": 2048, 00:31:40.334 "data_size": 63488 00:31:40.334 }, 00:31:40.334 { 00:31:40.334 "name": "BaseBdev4", 00:31:40.334 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:40.334 "is_configured": true, 00:31:40.334 "data_offset": 2048, 00:31:40.334 "data_size": 63488 00:31:40.334 } 00:31:40.334 ] 00:31:40.334 }' 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:40.334 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:40.900 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:31:40.900 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:40.900 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:40.900 [2024-10-01 20:30:35.943374] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:31:40.900 [2024-10-01 20:30:35.943508] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:40.900 [2024-10-01 20:30:35.943573] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c380 00:31:40.900 [2024-10-01 20:30:35.943613] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:40.900 [2024-10-01 20:30:35.944351] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:40.900 [2024-10-01 20:30:35.944408] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:31:40.900 [2024-10-01 20:30:35.944611] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:31:40.900 [2024-10-01 20:30:35.944665] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:31:40.900 [2024-10-01 20:30:35.944692] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:31:40.900 [2024-10-01 20:30:35.944778] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:31:40.900 [2024-10-01 20:30:35.957845] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000049370 00:31:40.900 spare 00:31:40.900 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:40.900 20:30:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:31:40.900 [2024-10-01 20:30:35.967464] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:31:41.832 20:30:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:31:41.833 20:30:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:41.833 20:30:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:31:41.833 20:30:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:31:41.833 20:30:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:41.833 20:30:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:41.833 20:30:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:41.833 20:30:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:41.833 20:30:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:41.833 20:30:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:41.833 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:41.833 "name": "raid_bdev1", 00:31:41.833 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:41.833 "strip_size_kb": 64, 00:31:41.833 "state": "online", 00:31:41.833 "raid_level": "raid5f", 00:31:41.833 "superblock": true, 00:31:41.833 "num_base_bdevs": 4, 00:31:41.833 "num_base_bdevs_discovered": 4, 00:31:41.833 "num_base_bdevs_operational": 4, 00:31:41.833 "process": { 00:31:41.833 "type": "rebuild", 00:31:41.833 "target": "spare", 00:31:41.833 "progress": { 00:31:41.833 "blocks": 17280, 00:31:41.833 "percent": 9 00:31:41.833 } 00:31:41.833 }, 00:31:41.833 "base_bdevs_list": [ 00:31:41.833 { 00:31:41.833 "name": "spare", 00:31:41.833 "uuid": "6c6280da-4ab3-5fad-a4b3-2467a9f6bfdc", 00:31:41.833 "is_configured": true, 00:31:41.833 "data_offset": 2048, 00:31:41.833 "data_size": 63488 00:31:41.833 }, 00:31:41.833 { 00:31:41.833 "name": "BaseBdev2", 00:31:41.833 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:41.833 "is_configured": true, 00:31:41.833 "data_offset": 2048, 00:31:41.833 "data_size": 63488 00:31:41.833 }, 00:31:41.833 { 00:31:41.833 "name": "BaseBdev3", 00:31:41.833 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:41.833 "is_configured": true, 00:31:41.833 "data_offset": 2048, 00:31:41.833 "data_size": 63488 00:31:41.833 }, 00:31:41.833 { 00:31:41.833 "name": "BaseBdev4", 00:31:41.833 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:41.833 "is_configured": true, 00:31:41.833 "data_offset": 2048, 00:31:41.833 "data_size": 63488 00:31:41.833 } 00:31:41.833 ] 00:31:41.833 }' 00:31:41.833 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:41.833 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:31:41.833 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:42.091 [2024-10-01 20:30:37.125115] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:42.091 [2024-10-01 20:30:37.180522] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:31:42.091 [2024-10-01 20:30:37.180646] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:42.091 [2024-10-01 20:30:37.180699] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:31:42.091 [2024-10-01 20:30:37.180744] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:42.091 "name": "raid_bdev1", 00:31:42.091 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:42.091 "strip_size_kb": 64, 00:31:42.091 "state": "online", 00:31:42.091 "raid_level": "raid5f", 00:31:42.091 "superblock": true, 00:31:42.091 "num_base_bdevs": 4, 00:31:42.091 "num_base_bdevs_discovered": 3, 00:31:42.091 "num_base_bdevs_operational": 3, 00:31:42.091 "base_bdevs_list": [ 00:31:42.091 { 00:31:42.091 "name": null, 00:31:42.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:42.091 "is_configured": false, 00:31:42.091 "data_offset": 0, 00:31:42.091 "data_size": 63488 00:31:42.091 }, 00:31:42.091 { 00:31:42.091 "name": "BaseBdev2", 00:31:42.091 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:42.091 "is_configured": true, 00:31:42.091 "data_offset": 2048, 00:31:42.091 "data_size": 63488 00:31:42.091 }, 00:31:42.091 { 00:31:42.091 "name": "BaseBdev3", 00:31:42.091 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:42.091 "is_configured": true, 00:31:42.091 "data_offset": 2048, 00:31:42.091 "data_size": 63488 00:31:42.091 }, 00:31:42.091 { 00:31:42.091 "name": "BaseBdev4", 00:31:42.091 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:42.091 "is_configured": true, 00:31:42.091 "data_offset": 2048, 00:31:42.091 "data_size": 63488 00:31:42.091 } 00:31:42.091 ] 00:31:42.091 }' 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:42.091 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:42.657 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:42.657 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:42.657 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:42.657 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:42.657 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:42.657 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:42.657 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:42.657 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:42.657 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:42.657 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:42.657 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:42.657 "name": "raid_bdev1", 00:31:42.657 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:42.657 "strip_size_kb": 64, 00:31:42.657 "state": "online", 00:31:42.657 "raid_level": "raid5f", 00:31:42.657 "superblock": true, 00:31:42.657 "num_base_bdevs": 4, 00:31:42.657 "num_base_bdevs_discovered": 3, 00:31:42.657 "num_base_bdevs_operational": 3, 00:31:42.657 "base_bdevs_list": [ 00:31:42.657 { 00:31:42.657 "name": null, 00:31:42.657 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:42.657 "is_configured": false, 00:31:42.657 "data_offset": 0, 00:31:42.657 "data_size": 63488 00:31:42.657 }, 00:31:42.657 { 00:31:42.657 "name": "BaseBdev2", 00:31:42.657 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:42.657 "is_configured": true, 00:31:42.657 "data_offset": 2048, 00:31:42.657 "data_size": 63488 00:31:42.657 }, 00:31:42.657 { 00:31:42.657 "name": "BaseBdev3", 00:31:42.657 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:42.657 "is_configured": true, 00:31:42.657 "data_offset": 2048, 00:31:42.657 "data_size": 63488 00:31:42.657 }, 00:31:42.657 { 00:31:42.657 "name": "BaseBdev4", 00:31:42.657 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:42.657 "is_configured": true, 00:31:42.657 "data_offset": 2048, 00:31:42.657 "data_size": 63488 00:31:42.657 } 00:31:42.657 ] 00:31:42.657 }' 00:31:42.657 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:42.657 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:42.657 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:42.657 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:42.657 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:31:42.658 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:42.658 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:42.658 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:42.658 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:31:42.658 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:42.658 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:42.658 [2024-10-01 20:30:37.903018] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:31:42.658 [2024-10-01 20:30:37.903166] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:42.658 [2024-10-01 20:30:37.903216] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c980 00:31:42.658 [2024-10-01 20:30:37.903233] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:42.658 [2024-10-01 20:30:37.903931] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:42.658 [2024-10-01 20:30:37.903971] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:31:42.658 [2024-10-01 20:30:37.904083] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:31:42.658 [2024-10-01 20:30:37.904117] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:31:42.658 [2024-10-01 20:30:37.904134] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:31:42.658 [2024-10-01 20:30:37.904149] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:31:42.658 BaseBdev1 00:31:42.658 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:42.658 20:30:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:44.031 "name": "raid_bdev1", 00:31:44.031 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:44.031 "strip_size_kb": 64, 00:31:44.031 "state": "online", 00:31:44.031 "raid_level": "raid5f", 00:31:44.031 "superblock": true, 00:31:44.031 "num_base_bdevs": 4, 00:31:44.031 "num_base_bdevs_discovered": 3, 00:31:44.031 "num_base_bdevs_operational": 3, 00:31:44.031 "base_bdevs_list": [ 00:31:44.031 { 00:31:44.031 "name": null, 00:31:44.031 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:44.031 "is_configured": false, 00:31:44.031 "data_offset": 0, 00:31:44.031 "data_size": 63488 00:31:44.031 }, 00:31:44.031 { 00:31:44.031 "name": "BaseBdev2", 00:31:44.031 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:44.031 "is_configured": true, 00:31:44.031 "data_offset": 2048, 00:31:44.031 "data_size": 63488 00:31:44.031 }, 00:31:44.031 { 00:31:44.031 "name": "BaseBdev3", 00:31:44.031 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:44.031 "is_configured": true, 00:31:44.031 "data_offset": 2048, 00:31:44.031 "data_size": 63488 00:31:44.031 }, 00:31:44.031 { 00:31:44.031 "name": "BaseBdev4", 00:31:44.031 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:44.031 "is_configured": true, 00:31:44.031 "data_offset": 2048, 00:31:44.031 "data_size": 63488 00:31:44.031 } 00:31:44.031 ] 00:31:44.031 }' 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:44.031 20:30:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:44.289 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:44.289 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:44.289 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:44.289 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:44.289 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:44.289 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:44.289 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:44.289 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:44.289 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:44.289 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:44.289 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:44.289 "name": "raid_bdev1", 00:31:44.289 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:44.289 "strip_size_kb": 64, 00:31:44.289 "state": "online", 00:31:44.289 "raid_level": "raid5f", 00:31:44.289 "superblock": true, 00:31:44.289 "num_base_bdevs": 4, 00:31:44.289 "num_base_bdevs_discovered": 3, 00:31:44.289 "num_base_bdevs_operational": 3, 00:31:44.289 "base_bdevs_list": [ 00:31:44.289 { 00:31:44.289 "name": null, 00:31:44.289 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:44.289 "is_configured": false, 00:31:44.289 "data_offset": 0, 00:31:44.289 "data_size": 63488 00:31:44.289 }, 00:31:44.289 { 00:31:44.289 "name": "BaseBdev2", 00:31:44.289 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:44.289 "is_configured": true, 00:31:44.289 "data_offset": 2048, 00:31:44.289 "data_size": 63488 00:31:44.289 }, 00:31:44.289 { 00:31:44.289 "name": "BaseBdev3", 00:31:44.289 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:44.289 "is_configured": true, 00:31:44.289 "data_offset": 2048, 00:31:44.289 "data_size": 63488 00:31:44.289 }, 00:31:44.289 { 00:31:44.289 "name": "BaseBdev4", 00:31:44.289 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:44.289 "is_configured": true, 00:31:44.289 "data_offset": 2048, 00:31:44.289 "data_size": 63488 00:31:44.289 } 00:31:44.289 ] 00:31:44.289 }' 00:31:44.289 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:44.548 [2024-10-01 20:30:39.619816] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:31:44.548 [2024-10-01 20:30:39.620079] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:31:44.548 [2024-10-01 20:30:39.620121] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:31:44.548 request: 00:31:44.548 { 00:31:44.548 "base_bdev": "BaseBdev1", 00:31:44.548 "raid_bdev": "raid_bdev1", 00:31:44.548 "method": "bdev_raid_add_base_bdev", 00:31:44.548 "req_id": 1 00:31:44.548 } 00:31:44.548 Got JSON-RPC error response 00:31:44.548 response: 00:31:44.548 { 00:31:44.548 "code": -22, 00:31:44.548 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:31:44.548 } 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:31:44.548 20:30:39 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:45.482 "name": "raid_bdev1", 00:31:45.482 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:45.482 "strip_size_kb": 64, 00:31:45.482 "state": "online", 00:31:45.482 "raid_level": "raid5f", 00:31:45.482 "superblock": true, 00:31:45.482 "num_base_bdevs": 4, 00:31:45.482 "num_base_bdevs_discovered": 3, 00:31:45.482 "num_base_bdevs_operational": 3, 00:31:45.482 "base_bdevs_list": [ 00:31:45.482 { 00:31:45.482 "name": null, 00:31:45.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:45.482 "is_configured": false, 00:31:45.482 "data_offset": 0, 00:31:45.482 "data_size": 63488 00:31:45.482 }, 00:31:45.482 { 00:31:45.482 "name": "BaseBdev2", 00:31:45.482 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:45.482 "is_configured": true, 00:31:45.482 "data_offset": 2048, 00:31:45.482 "data_size": 63488 00:31:45.482 }, 00:31:45.482 { 00:31:45.482 "name": "BaseBdev3", 00:31:45.482 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:45.482 "is_configured": true, 00:31:45.482 "data_offset": 2048, 00:31:45.482 "data_size": 63488 00:31:45.482 }, 00:31:45.482 { 00:31:45.482 "name": "BaseBdev4", 00:31:45.482 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:45.482 "is_configured": true, 00:31:45.482 "data_offset": 2048, 00:31:45.482 "data_size": 63488 00:31:45.482 } 00:31:45.482 ] 00:31:45.482 }' 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:45.482 20:30:40 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:46.048 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:31:46.048 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:31:46.048 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:31:46.048 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:31:46.048 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:31:46.048 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:46.048 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:46.048 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:46.048 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:46.048 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:46.048 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:31:46.048 "name": "raid_bdev1", 00:31:46.048 "uuid": "d674d811-1970-4936-8544-aecf58c11db2", 00:31:46.048 "strip_size_kb": 64, 00:31:46.048 "state": "online", 00:31:46.048 "raid_level": "raid5f", 00:31:46.048 "superblock": true, 00:31:46.048 "num_base_bdevs": 4, 00:31:46.048 "num_base_bdevs_discovered": 3, 00:31:46.048 "num_base_bdevs_operational": 3, 00:31:46.048 "base_bdevs_list": [ 00:31:46.048 { 00:31:46.048 "name": null, 00:31:46.048 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:46.048 "is_configured": false, 00:31:46.048 "data_offset": 0, 00:31:46.048 "data_size": 63488 00:31:46.048 }, 00:31:46.048 { 00:31:46.048 "name": "BaseBdev2", 00:31:46.048 "uuid": "0537b734-1edb-5c00-bf60-f485b8310d4c", 00:31:46.048 "is_configured": true, 00:31:46.048 "data_offset": 2048, 00:31:46.048 "data_size": 63488 00:31:46.048 }, 00:31:46.048 { 00:31:46.048 "name": "BaseBdev3", 00:31:46.048 "uuid": "4a22b213-ee7c-56bc-872c-79d323ef74f9", 00:31:46.048 "is_configured": true, 00:31:46.048 "data_offset": 2048, 00:31:46.048 "data_size": 63488 00:31:46.048 }, 00:31:46.048 { 00:31:46.048 "name": "BaseBdev4", 00:31:46.048 "uuid": "de699d00-d8ca-5636-9e8c-205a607ebd34", 00:31:46.048 "is_configured": true, 00:31:46.048 "data_offset": 2048, 00:31:46.048 "data_size": 63488 00:31:46.048 } 00:31:46.048 ] 00:31:46.048 }' 00:31:46.048 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:31:46.048 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:31:46.048 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:31:46.317 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:31:46.317 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 86488 00:31:46.317 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@950 -- # '[' -z 86488 ']' 00:31:46.317 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # kill -0 86488 00:31:46.317 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@955 -- # uname 00:31:46.317 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:31:46.317 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 86488 00:31:46.317 killing process with pid 86488 00:31:46.317 Received shutdown signal, test time was about 60.000000 seconds 00:31:46.317 00:31:46.317 Latency(us) 00:31:46.317 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:46.317 =================================================================================================================== 00:31:46.317 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:31:46.317 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:31:46.317 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:31:46.317 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 86488' 00:31:46.317 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@969 -- # kill 86488 00:31:46.317 [2024-10-01 20:30:41.338804] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:31:46.317 20:30:41 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@974 -- # wait 86488 00:31:46.317 [2024-10-01 20:30:41.338988] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:31:46.317 [2024-10-01 20:30:41.339095] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:31:46.317 [2024-10-01 20:30:41.339119] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:31:46.881 [2024-10-01 20:30:41.829142] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:31:48.790 20:30:43 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:31:48.790 00:31:48.790 real 0m30.666s 00:31:48.790 user 0m39.353s 00:31:48.790 sys 0m3.322s 00:31:48.790 20:30:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:31:48.790 ************************************ 00:31:48.790 20:30:43 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:31:48.790 END TEST raid5f_rebuild_test_sb 00:31:48.790 ************************************ 00:31:48.790 20:30:44 bdev_raid -- bdev/bdev_raid.sh@995 -- # base_blocklen=4096 00:31:48.790 20:30:44 bdev_raid -- bdev/bdev_raid.sh@997 -- # run_test raid_state_function_test_sb_4k raid_state_function_test raid1 2 true 00:31:48.790 20:30:44 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:31:48.790 20:30:44 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:31:48.790 20:30:44 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:31:49.049 ************************************ 00:31:49.049 START TEST raid_state_function_test_sb_4k 00:31:49.049 ************************************ 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 true 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # local strip_size 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@229 -- # raid_pid=87328 00:31:49.049 Process raid pid: 87328 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 87328' 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@231 -- # waitforlisten 87328 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@831 -- # '[' -z 87328 ']' 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@836 -- # local max_retries=100 00:31:49.049 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@840 -- # xtrace_disable 00:31:49.049 20:30:44 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:49.049 [2024-10-01 20:30:44.181965] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:31:49.049 [2024-10-01 20:30:44.182186] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:31:49.307 [2024-10-01 20:30:44.363941] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:49.565 [2024-10-01 20:30:44.663880] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:31:49.824 [2024-10-01 20:30:44.920048] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:31:49.824 [2024-10-01 20:30:44.920148] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@864 -- # return 0 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:50.390 [2024-10-01 20:30:45.421748] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:31:50.390 [2024-10-01 20:30:45.421819] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:31:50.390 [2024-10-01 20:30:45.421837] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:31:50.390 [2024-10-01 20:30:45.421855] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:50.390 "name": "Existed_Raid", 00:31:50.390 "uuid": "ea0dc5e4-5fb4-47e8-88e4-2268433e3126", 00:31:50.390 "strip_size_kb": 0, 00:31:50.390 "state": "configuring", 00:31:50.390 "raid_level": "raid1", 00:31:50.390 "superblock": true, 00:31:50.390 "num_base_bdevs": 2, 00:31:50.390 "num_base_bdevs_discovered": 0, 00:31:50.390 "num_base_bdevs_operational": 2, 00:31:50.390 "base_bdevs_list": [ 00:31:50.390 { 00:31:50.390 "name": "BaseBdev1", 00:31:50.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:50.390 "is_configured": false, 00:31:50.390 "data_offset": 0, 00:31:50.390 "data_size": 0 00:31:50.390 }, 00:31:50.390 { 00:31:50.390 "name": "BaseBdev2", 00:31:50.390 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:50.390 "is_configured": false, 00:31:50.390 "data_offset": 0, 00:31:50.390 "data_size": 0 00:31:50.390 } 00:31:50.390 ] 00:31:50.390 }' 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:50.390 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:50.956 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:31:50.956 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:50.956 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:50.956 [2024-10-01 20:30:45.969806] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:31:50.956 [2024-10-01 20:30:45.969866] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:31:50.957 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:50.957 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:31:50.957 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:50.957 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:50.957 [2024-10-01 20:30:45.981865] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:31:50.957 [2024-10-01 20:30:45.981997] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:31:50.957 [2024-10-01 20:30:45.982028] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:31:50.957 [2024-10-01 20:30:45.982065] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:31:50.957 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:50.957 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1 00:31:50.957 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:50.957 20:30:45 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:50.957 [2024-10-01 20:30:46.037534] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:31:50.957 BaseBdev1 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@901 -- # local i 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:50.957 [ 00:31:50.957 { 00:31:50.957 "name": "BaseBdev1", 00:31:50.957 "aliases": [ 00:31:50.957 "8d8b7b66-ac94-4355-97e5-d5f15a51d843" 00:31:50.957 ], 00:31:50.957 "product_name": "Malloc disk", 00:31:50.957 "block_size": 4096, 00:31:50.957 "num_blocks": 8192, 00:31:50.957 "uuid": "8d8b7b66-ac94-4355-97e5-d5f15a51d843", 00:31:50.957 "assigned_rate_limits": { 00:31:50.957 "rw_ios_per_sec": 0, 00:31:50.957 "rw_mbytes_per_sec": 0, 00:31:50.957 "r_mbytes_per_sec": 0, 00:31:50.957 "w_mbytes_per_sec": 0 00:31:50.957 }, 00:31:50.957 "claimed": true, 00:31:50.957 "claim_type": "exclusive_write", 00:31:50.957 "zoned": false, 00:31:50.957 "supported_io_types": { 00:31:50.957 "read": true, 00:31:50.957 "write": true, 00:31:50.957 "unmap": true, 00:31:50.957 "flush": true, 00:31:50.957 "reset": true, 00:31:50.957 "nvme_admin": false, 00:31:50.957 "nvme_io": false, 00:31:50.957 "nvme_io_md": false, 00:31:50.957 "write_zeroes": true, 00:31:50.957 "zcopy": true, 00:31:50.957 "get_zone_info": false, 00:31:50.957 "zone_management": false, 00:31:50.957 "zone_append": false, 00:31:50.957 "compare": false, 00:31:50.957 "compare_and_write": false, 00:31:50.957 "abort": true, 00:31:50.957 "seek_hole": false, 00:31:50.957 "seek_data": false, 00:31:50.957 "copy": true, 00:31:50.957 "nvme_iov_md": false 00:31:50.957 }, 00:31:50.957 "memory_domains": [ 00:31:50.957 { 00:31:50.957 "dma_device_id": "system", 00:31:50.957 "dma_device_type": 1 00:31:50.957 }, 00:31:50.957 { 00:31:50.957 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:31:50.957 "dma_device_type": 2 00:31:50.957 } 00:31:50.957 ], 00:31:50.957 "driver_specific": {} 00:31:50.957 } 00:31:50.957 ] 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@907 -- # return 0 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:50.957 "name": "Existed_Raid", 00:31:50.957 "uuid": "26d70b49-87ed-457e-9c2d-7d0f23574e40", 00:31:50.957 "strip_size_kb": 0, 00:31:50.957 "state": "configuring", 00:31:50.957 "raid_level": "raid1", 00:31:50.957 "superblock": true, 00:31:50.957 "num_base_bdevs": 2, 00:31:50.957 "num_base_bdevs_discovered": 1, 00:31:50.957 "num_base_bdevs_operational": 2, 00:31:50.957 "base_bdevs_list": [ 00:31:50.957 { 00:31:50.957 "name": "BaseBdev1", 00:31:50.957 "uuid": "8d8b7b66-ac94-4355-97e5-d5f15a51d843", 00:31:50.957 "is_configured": true, 00:31:50.957 "data_offset": 256, 00:31:50.957 "data_size": 7936 00:31:50.957 }, 00:31:50.957 { 00:31:50.957 "name": "BaseBdev2", 00:31:50.957 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:50.957 "is_configured": false, 00:31:50.957 "data_offset": 0, 00:31:50.957 "data_size": 0 00:31:50.957 } 00:31:50.957 ] 00:31:50.957 }' 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:50.957 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:51.523 [2024-10-01 20:30:46.605792] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:31:51.523 [2024-10-01 20:30:46.605872] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:51.523 [2024-10-01 20:30:46.613852] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:31:51.523 [2024-10-01 20:30:46.616623] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:31:51.523 [2024-10-01 20:30:46.616770] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:51.523 "name": "Existed_Raid", 00:31:51.523 "uuid": "86b5c3a3-d2eb-49e3-8ade-3e99618c474f", 00:31:51.523 "strip_size_kb": 0, 00:31:51.523 "state": "configuring", 00:31:51.523 "raid_level": "raid1", 00:31:51.523 "superblock": true, 00:31:51.523 "num_base_bdevs": 2, 00:31:51.523 "num_base_bdevs_discovered": 1, 00:31:51.523 "num_base_bdevs_operational": 2, 00:31:51.523 "base_bdevs_list": [ 00:31:51.523 { 00:31:51.523 "name": "BaseBdev1", 00:31:51.523 "uuid": "8d8b7b66-ac94-4355-97e5-d5f15a51d843", 00:31:51.523 "is_configured": true, 00:31:51.523 "data_offset": 256, 00:31:51.523 "data_size": 7936 00:31:51.523 }, 00:31:51.523 { 00:31:51.523 "name": "BaseBdev2", 00:31:51.523 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:51.523 "is_configured": false, 00:31:51.523 "data_offset": 0, 00:31:51.523 "data_size": 0 00:31:51.523 } 00:31:51.523 ] 00:31:51.523 }' 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:51.523 20:30:46 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:52.091 [2024-10-01 20:30:47.214417] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:31:52.091 [2024-10-01 20:30:47.214749] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:31:52.091 [2024-10-01 20:30:47.214776] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:31:52.091 [2024-10-01 20:30:47.215114] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:31:52.091 BaseBdev2 00:31:52.091 [2024-10-01 20:30:47.215324] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:31:52.091 [2024-10-01 20:30:47.215348] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:31:52.091 [2024-10-01 20:30:47.215557] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@901 -- # local i 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:52.091 [ 00:31:52.091 { 00:31:52.091 "name": "BaseBdev2", 00:31:52.091 "aliases": [ 00:31:52.091 "4915b5e3-eff4-4597-a6e5-727403e97319" 00:31:52.091 ], 00:31:52.091 "product_name": "Malloc disk", 00:31:52.091 "block_size": 4096, 00:31:52.091 "num_blocks": 8192, 00:31:52.091 "uuid": "4915b5e3-eff4-4597-a6e5-727403e97319", 00:31:52.091 "assigned_rate_limits": { 00:31:52.091 "rw_ios_per_sec": 0, 00:31:52.091 "rw_mbytes_per_sec": 0, 00:31:52.091 "r_mbytes_per_sec": 0, 00:31:52.091 "w_mbytes_per_sec": 0 00:31:52.091 }, 00:31:52.091 "claimed": true, 00:31:52.091 "claim_type": "exclusive_write", 00:31:52.091 "zoned": false, 00:31:52.091 "supported_io_types": { 00:31:52.091 "read": true, 00:31:52.091 "write": true, 00:31:52.091 "unmap": true, 00:31:52.091 "flush": true, 00:31:52.091 "reset": true, 00:31:52.091 "nvme_admin": false, 00:31:52.091 "nvme_io": false, 00:31:52.091 "nvme_io_md": false, 00:31:52.091 "write_zeroes": true, 00:31:52.091 "zcopy": true, 00:31:52.091 "get_zone_info": false, 00:31:52.091 "zone_management": false, 00:31:52.091 "zone_append": false, 00:31:52.091 "compare": false, 00:31:52.091 "compare_and_write": false, 00:31:52.091 "abort": true, 00:31:52.091 "seek_hole": false, 00:31:52.091 "seek_data": false, 00:31:52.091 "copy": true, 00:31:52.091 "nvme_iov_md": false 00:31:52.091 }, 00:31:52.091 "memory_domains": [ 00:31:52.091 { 00:31:52.091 "dma_device_id": "system", 00:31:52.091 "dma_device_type": 1 00:31:52.091 }, 00:31:52.091 { 00:31:52.091 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:31:52.091 "dma_device_type": 2 00:31:52.091 } 00:31:52.091 ], 00:31:52.091 "driver_specific": {} 00:31:52.091 } 00:31:52.091 ] 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@907 -- # return 0 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:52.091 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:52.091 "name": "Existed_Raid", 00:31:52.091 "uuid": "86b5c3a3-d2eb-49e3-8ade-3e99618c474f", 00:31:52.091 "strip_size_kb": 0, 00:31:52.091 "state": "online", 00:31:52.091 "raid_level": "raid1", 00:31:52.091 "superblock": true, 00:31:52.091 "num_base_bdevs": 2, 00:31:52.091 "num_base_bdevs_discovered": 2, 00:31:52.091 "num_base_bdevs_operational": 2, 00:31:52.091 "base_bdevs_list": [ 00:31:52.091 { 00:31:52.091 "name": "BaseBdev1", 00:31:52.092 "uuid": "8d8b7b66-ac94-4355-97e5-d5f15a51d843", 00:31:52.092 "is_configured": true, 00:31:52.092 "data_offset": 256, 00:31:52.092 "data_size": 7936 00:31:52.092 }, 00:31:52.092 { 00:31:52.092 "name": "BaseBdev2", 00:31:52.092 "uuid": "4915b5e3-eff4-4597-a6e5-727403e97319", 00:31:52.092 "is_configured": true, 00:31:52.092 "data_offset": 256, 00:31:52.092 "data_size": 7936 00:31:52.092 } 00:31:52.092 ] 00:31:52.092 }' 00:31:52.092 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:52.092 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:52.659 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:31:52.659 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:31:52.659 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:31:52.659 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:31:52.659 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local name 00:31:52.659 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:31:52.659 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:31:52.659 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:31:52.659 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:52.659 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:52.659 [2024-10-01 20:30:47.787240] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:31:52.659 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:52.659 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:31:52.659 "name": "Existed_Raid", 00:31:52.659 "aliases": [ 00:31:52.659 "86b5c3a3-d2eb-49e3-8ade-3e99618c474f" 00:31:52.659 ], 00:31:52.659 "product_name": "Raid Volume", 00:31:52.659 "block_size": 4096, 00:31:52.659 "num_blocks": 7936, 00:31:52.659 "uuid": "86b5c3a3-d2eb-49e3-8ade-3e99618c474f", 00:31:52.659 "assigned_rate_limits": { 00:31:52.659 "rw_ios_per_sec": 0, 00:31:52.659 "rw_mbytes_per_sec": 0, 00:31:52.659 "r_mbytes_per_sec": 0, 00:31:52.659 "w_mbytes_per_sec": 0 00:31:52.659 }, 00:31:52.659 "claimed": false, 00:31:52.659 "zoned": false, 00:31:52.659 "supported_io_types": { 00:31:52.659 "read": true, 00:31:52.659 "write": true, 00:31:52.659 "unmap": false, 00:31:52.659 "flush": false, 00:31:52.659 "reset": true, 00:31:52.659 "nvme_admin": false, 00:31:52.659 "nvme_io": false, 00:31:52.659 "nvme_io_md": false, 00:31:52.659 "write_zeroes": true, 00:31:52.659 "zcopy": false, 00:31:52.659 "get_zone_info": false, 00:31:52.659 "zone_management": false, 00:31:52.659 "zone_append": false, 00:31:52.659 "compare": false, 00:31:52.659 "compare_and_write": false, 00:31:52.659 "abort": false, 00:31:52.659 "seek_hole": false, 00:31:52.659 "seek_data": false, 00:31:52.659 "copy": false, 00:31:52.659 "nvme_iov_md": false 00:31:52.659 }, 00:31:52.659 "memory_domains": [ 00:31:52.659 { 00:31:52.659 "dma_device_id": "system", 00:31:52.659 "dma_device_type": 1 00:31:52.659 }, 00:31:52.659 { 00:31:52.659 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:31:52.659 "dma_device_type": 2 00:31:52.659 }, 00:31:52.659 { 00:31:52.659 "dma_device_id": "system", 00:31:52.659 "dma_device_type": 1 00:31:52.659 }, 00:31:52.659 { 00:31:52.659 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:31:52.659 "dma_device_type": 2 00:31:52.659 } 00:31:52.659 ], 00:31:52.659 "driver_specific": { 00:31:52.659 "raid": { 00:31:52.659 "uuid": "86b5c3a3-d2eb-49e3-8ade-3e99618c474f", 00:31:52.659 "strip_size_kb": 0, 00:31:52.659 "state": "online", 00:31:52.659 "raid_level": "raid1", 00:31:52.659 "superblock": true, 00:31:52.659 "num_base_bdevs": 2, 00:31:52.659 "num_base_bdevs_discovered": 2, 00:31:52.659 "num_base_bdevs_operational": 2, 00:31:52.659 "base_bdevs_list": [ 00:31:52.659 { 00:31:52.659 "name": "BaseBdev1", 00:31:52.659 "uuid": "8d8b7b66-ac94-4355-97e5-d5f15a51d843", 00:31:52.659 "is_configured": true, 00:31:52.659 "data_offset": 256, 00:31:52.659 "data_size": 7936 00:31:52.659 }, 00:31:52.659 { 00:31:52.659 "name": "BaseBdev2", 00:31:52.659 "uuid": "4915b5e3-eff4-4597-a6e5-727403e97319", 00:31:52.659 "is_configured": true, 00:31:52.659 "data_offset": 256, 00:31:52.659 "data_size": 7936 00:31:52.659 } 00:31:52.659 ] 00:31:52.659 } 00:31:52.659 } 00:31:52.659 }' 00:31:52.659 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:31:52.659 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:31:52.659 BaseBdev2' 00:31:52.659 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:31:52.919 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:31:52.919 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:31:52.919 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:31:52.919 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:52.919 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:52.919 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:31:52.919 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:52.919 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:31:52.919 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:31:52.919 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:31:52.919 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:31:52.919 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:31:52.919 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:52.919 20:30:47 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:52.919 [2024-10-01 20:30:48.055025] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@260 -- # local expected_state 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:31:52.919 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:53.179 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:53.179 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:53.179 "name": "Existed_Raid", 00:31:53.179 "uuid": "86b5c3a3-d2eb-49e3-8ade-3e99618c474f", 00:31:53.179 "strip_size_kb": 0, 00:31:53.179 "state": "online", 00:31:53.179 "raid_level": "raid1", 00:31:53.179 "superblock": true, 00:31:53.179 "num_base_bdevs": 2, 00:31:53.179 "num_base_bdevs_discovered": 1, 00:31:53.179 "num_base_bdevs_operational": 1, 00:31:53.179 "base_bdevs_list": [ 00:31:53.179 { 00:31:53.179 "name": null, 00:31:53.179 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:53.179 "is_configured": false, 00:31:53.179 "data_offset": 0, 00:31:53.179 "data_size": 7936 00:31:53.179 }, 00:31:53.179 { 00:31:53.179 "name": "BaseBdev2", 00:31:53.179 "uuid": "4915b5e3-eff4-4597-a6e5-727403e97319", 00:31:53.179 "is_configured": true, 00:31:53.179 "data_offset": 256, 00:31:53.179 "data_size": 7936 00:31:53.179 } 00:31:53.179 ] 00:31:53.179 }' 00:31:53.179 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:53.179 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:53.439 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:53.699 [2024-10-01 20:30:48.750143] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:31:53.699 [2024-10-01 20:30:48.750290] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:31:53.699 [2024-10-01 20:30:48.848474] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:31:53.699 [2024-10-01 20:30:48.848549] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:31:53.699 [2024-10-01 20:30:48.848571] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@326 -- # killprocess 87328 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@950 -- # '[' -z 87328 ']' 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@954 -- # kill -0 87328 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@955 -- # uname 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 87328 00:31:53.699 killing process with pid 87328 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@968 -- # echo 'killing process with pid 87328' 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@969 -- # kill 87328 00:31:53.699 [2024-10-01 20:30:48.938696] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:31:53.699 20:30:48 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@974 -- # wait 87328 00:31:53.958 [2024-10-01 20:30:48.954761] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:31:55.863 ************************************ 00:31:55.863 END TEST raid_state_function_test_sb_4k 00:31:55.863 ************************************ 00:31:55.863 20:30:50 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@328 -- # return 0 00:31:55.863 00:31:55.863 real 0m6.898s 00:31:55.863 user 0m9.672s 00:31:55.863 sys 0m1.104s 00:31:55.863 20:30:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1126 -- # xtrace_disable 00:31:55.863 20:30:50 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:31:55.863 20:30:51 bdev_raid -- bdev/bdev_raid.sh@998 -- # run_test raid_superblock_test_4k raid_superblock_test raid1 2 00:31:55.863 20:30:51 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:31:55.863 20:30:51 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:31:55.863 20:30:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:31:55.863 ************************************ 00:31:55.863 START TEST raid_superblock_test_4k 00:31:55.863 ************************************ 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 2 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@399 -- # local strip_size 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:31:55.863 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@412 -- # raid_pid=87597 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@413 -- # waitforlisten 87597 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@831 -- # '[' -z 87597 ']' 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@836 -- # local max_retries=100 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@840 -- # xtrace_disable 00:31:55.863 20:30:51 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:55.863 [2024-10-01 20:30:51.104660] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:31:55.863 [2024-10-01 20:30:51.105026] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87597 ] 00:31:56.123 [2024-10-01 20:30:51.263841] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:56.383 [2024-10-01 20:30:51.516349] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:31:56.643 [2024-10-01 20:30:51.727374] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:31:56.643 [2024-10-01 20:30:51.727618] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@864 -- # return 0 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc1 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:56.901 malloc1 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:56.901 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:57.160 [2024-10-01 20:30:52.157710] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:31:57.160 [2024-10-01 20:30:52.158040] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:57.160 [2024-10-01 20:30:52.158140] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:31:57.160 [2024-10-01 20:30:52.158167] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:57.160 [2024-10-01 20:30:52.161189] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:57.160 [2024-10-01 20:30:52.161486] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:31:57.160 pt1 00:31:57.160 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:57.160 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:31:57.160 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:31:57.160 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:31:57.160 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:31:57.160 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:31:57.160 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:31:57.160 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:31:57.160 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:31:57.160 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc2 00:31:57.160 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:57.160 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:57.160 malloc2 00:31:57.160 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:57.160 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:31:57.160 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:57.160 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:57.160 [2024-10-01 20:30:52.210352] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:31:57.160 [2024-10-01 20:30:52.210435] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:57.160 [2024-10-01 20:30:52.210472] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:31:57.161 [2024-10-01 20:30:52.210487] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:57.161 [2024-10-01 20:30:52.213423] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:57.161 [2024-10-01 20:30:52.213466] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:31:57.161 pt2 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:57.161 [2024-10-01 20:30:52.222463] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:31:57.161 [2024-10-01 20:30:52.225032] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:31:57.161 [2024-10-01 20:30:52.225499] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:31:57.161 [2024-10-01 20:30:52.225526] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:31:57.161 [2024-10-01 20:30:52.225933] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:31:57.161 [2024-10-01 20:30:52.226142] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:31:57.161 [2024-10-01 20:30:52.226212] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:31:57.161 [2024-10-01 20:30:52.226476] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:57.161 "name": "raid_bdev1", 00:31:57.161 "uuid": "e5061f8a-8b03-443a-aaf1-9252fe7ee9a2", 00:31:57.161 "strip_size_kb": 0, 00:31:57.161 "state": "online", 00:31:57.161 "raid_level": "raid1", 00:31:57.161 "superblock": true, 00:31:57.161 "num_base_bdevs": 2, 00:31:57.161 "num_base_bdevs_discovered": 2, 00:31:57.161 "num_base_bdevs_operational": 2, 00:31:57.161 "base_bdevs_list": [ 00:31:57.161 { 00:31:57.161 "name": "pt1", 00:31:57.161 "uuid": "00000000-0000-0000-0000-000000000001", 00:31:57.161 "is_configured": true, 00:31:57.161 "data_offset": 256, 00:31:57.161 "data_size": 7936 00:31:57.161 }, 00:31:57.161 { 00:31:57.161 "name": "pt2", 00:31:57.161 "uuid": "00000000-0000-0000-0000-000000000002", 00:31:57.161 "is_configured": true, 00:31:57.161 "data_offset": 256, 00:31:57.161 "data_size": 7936 00:31:57.161 } 00:31:57.161 ] 00:31:57.161 }' 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:57.161 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:57.728 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:31:57.728 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:31:57.728 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:31:57.728 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:31:57.728 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:31:57.728 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:31:57.728 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:31:57.728 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:57.728 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:31:57.728 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:57.728 [2024-10-01 20:30:52.763128] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:31:57.728 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:57.728 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:31:57.728 "name": "raid_bdev1", 00:31:57.728 "aliases": [ 00:31:57.728 "e5061f8a-8b03-443a-aaf1-9252fe7ee9a2" 00:31:57.728 ], 00:31:57.728 "product_name": "Raid Volume", 00:31:57.728 "block_size": 4096, 00:31:57.728 "num_blocks": 7936, 00:31:57.728 "uuid": "e5061f8a-8b03-443a-aaf1-9252fe7ee9a2", 00:31:57.728 "assigned_rate_limits": { 00:31:57.728 "rw_ios_per_sec": 0, 00:31:57.728 "rw_mbytes_per_sec": 0, 00:31:57.728 "r_mbytes_per_sec": 0, 00:31:57.728 "w_mbytes_per_sec": 0 00:31:57.728 }, 00:31:57.728 "claimed": false, 00:31:57.728 "zoned": false, 00:31:57.728 "supported_io_types": { 00:31:57.728 "read": true, 00:31:57.728 "write": true, 00:31:57.728 "unmap": false, 00:31:57.728 "flush": false, 00:31:57.728 "reset": true, 00:31:57.728 "nvme_admin": false, 00:31:57.728 "nvme_io": false, 00:31:57.728 "nvme_io_md": false, 00:31:57.728 "write_zeroes": true, 00:31:57.728 "zcopy": false, 00:31:57.728 "get_zone_info": false, 00:31:57.728 "zone_management": false, 00:31:57.728 "zone_append": false, 00:31:57.728 "compare": false, 00:31:57.728 "compare_and_write": false, 00:31:57.728 "abort": false, 00:31:57.728 "seek_hole": false, 00:31:57.728 "seek_data": false, 00:31:57.728 "copy": false, 00:31:57.728 "nvme_iov_md": false 00:31:57.728 }, 00:31:57.728 "memory_domains": [ 00:31:57.728 { 00:31:57.728 "dma_device_id": "system", 00:31:57.728 "dma_device_type": 1 00:31:57.728 }, 00:31:57.728 { 00:31:57.728 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:31:57.728 "dma_device_type": 2 00:31:57.728 }, 00:31:57.728 { 00:31:57.728 "dma_device_id": "system", 00:31:57.728 "dma_device_type": 1 00:31:57.728 }, 00:31:57.728 { 00:31:57.728 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:31:57.728 "dma_device_type": 2 00:31:57.728 } 00:31:57.728 ], 00:31:57.728 "driver_specific": { 00:31:57.728 "raid": { 00:31:57.728 "uuid": "e5061f8a-8b03-443a-aaf1-9252fe7ee9a2", 00:31:57.728 "strip_size_kb": 0, 00:31:57.728 "state": "online", 00:31:57.728 "raid_level": "raid1", 00:31:57.728 "superblock": true, 00:31:57.728 "num_base_bdevs": 2, 00:31:57.728 "num_base_bdevs_discovered": 2, 00:31:57.728 "num_base_bdevs_operational": 2, 00:31:57.729 "base_bdevs_list": [ 00:31:57.729 { 00:31:57.729 "name": "pt1", 00:31:57.729 "uuid": "00000000-0000-0000-0000-000000000001", 00:31:57.729 "is_configured": true, 00:31:57.729 "data_offset": 256, 00:31:57.729 "data_size": 7936 00:31:57.729 }, 00:31:57.729 { 00:31:57.729 "name": "pt2", 00:31:57.729 "uuid": "00000000-0000-0000-0000-000000000002", 00:31:57.729 "is_configured": true, 00:31:57.729 "data_offset": 256, 00:31:57.729 "data_size": 7936 00:31:57.729 } 00:31:57.729 ] 00:31:57.729 } 00:31:57.729 } 00:31:57.729 }' 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:31:57.729 pt2' 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:31:57.729 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:57.988 20:30:52 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:57.988 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:31:57.988 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:31:57.988 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:31:57.988 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:31:57.988 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:57.988 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:57.988 [2024-10-01 20:30:53.035010] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:31:57.988 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:57.988 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=e5061f8a-8b03-443a-aaf1-9252fe7ee9a2 00:31:57.988 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@436 -- # '[' -z e5061f8a-8b03-443a-aaf1-9252fe7ee9a2 ']' 00:31:57.988 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:31:57.988 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:57.988 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:57.989 [2024-10-01 20:30:53.082685] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:31:57.989 [2024-10-01 20:30:53.082712] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:31:57.989 [2024-10-01 20:30:53.082857] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:31:57.989 [2024-10-01 20:30:53.082940] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:31:57.989 [2024-10-01 20:30:53.082961] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@650 -- # local es=0 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:57.989 [2024-10-01 20:30:53.230833] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:31:57.989 [2024-10-01 20:30:53.233651] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:31:57.989 [2024-10-01 20:30:53.233788] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:31:57.989 [2024-10-01 20:30:53.233889] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:31:57.989 [2024-10-01 20:30:53.233918] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:31:57.989 [2024-10-01 20:30:53.233934] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:31:57.989 request: 00:31:57.989 { 00:31:57.989 "name": "raid_bdev1", 00:31:57.989 "raid_level": "raid1", 00:31:57.989 "base_bdevs": [ 00:31:57.989 "malloc1", 00:31:57.989 "malloc2" 00:31:57.989 ], 00:31:57.989 "superblock": false, 00:31:57.989 "method": "bdev_raid_create", 00:31:57.989 "req_id": 1 00:31:57.989 } 00:31:57.989 Got JSON-RPC error response 00:31:57.989 response: 00:31:57.989 { 00:31:57.989 "code": -17, 00:31:57.989 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:31:57.989 } 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@653 -- # es=1 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:31:57.989 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:31:58.248 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:58.248 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:31:58.248 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:58.248 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:58.248 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:58.248 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:31:58.248 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:31:58.248 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:31:58.248 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:58.248 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:58.248 [2024-10-01 20:30:53.294888] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:31:58.248 [2024-10-01 20:30:53.295151] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:58.248 [2024-10-01 20:30:53.295288] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:31:58.248 [2024-10-01 20:30:53.295404] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:58.248 [2024-10-01 20:30:53.298489] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:58.248 [2024-10-01 20:30:53.298676] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:31:58.248 [2024-10-01 20:30:53.298957] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:31:58.248 [2024-10-01 20:30:53.299172] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:31:58.248 pt1 00:31:58.248 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:58.249 "name": "raid_bdev1", 00:31:58.249 "uuid": "e5061f8a-8b03-443a-aaf1-9252fe7ee9a2", 00:31:58.249 "strip_size_kb": 0, 00:31:58.249 "state": "configuring", 00:31:58.249 "raid_level": "raid1", 00:31:58.249 "superblock": true, 00:31:58.249 "num_base_bdevs": 2, 00:31:58.249 "num_base_bdevs_discovered": 1, 00:31:58.249 "num_base_bdevs_operational": 2, 00:31:58.249 "base_bdevs_list": [ 00:31:58.249 { 00:31:58.249 "name": "pt1", 00:31:58.249 "uuid": "00000000-0000-0000-0000-000000000001", 00:31:58.249 "is_configured": true, 00:31:58.249 "data_offset": 256, 00:31:58.249 "data_size": 7936 00:31:58.249 }, 00:31:58.249 { 00:31:58.249 "name": null, 00:31:58.249 "uuid": "00000000-0000-0000-0000-000000000002", 00:31:58.249 "is_configured": false, 00:31:58.249 "data_offset": 256, 00:31:58.249 "data_size": 7936 00:31:58.249 } 00:31:58.249 ] 00:31:58.249 }' 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:58.249 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:58.816 [2024-10-01 20:30:53.847214] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:31:58.816 [2024-10-01 20:30:53.847352] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:31:58.816 [2024-10-01 20:30:53.847420] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:31:58.816 [2024-10-01 20:30:53.847439] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:31:58.816 [2024-10-01 20:30:53.848241] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:31:58.816 [2024-10-01 20:30:53.848420] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:31:58.816 [2024-10-01 20:30:53.848562] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:31:58.816 [2024-10-01 20:30:53.848603] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:31:58.816 [2024-10-01 20:30:53.848855] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:31:58.816 [2024-10-01 20:30:53.848880] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:31:58.816 [2024-10-01 20:30:53.849249] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:31:58.816 [2024-10-01 20:30:53.849509] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:31:58.816 [2024-10-01 20:30:53.849527] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:31:58.816 [2024-10-01 20:30:53.849728] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:31:58.816 pt2 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:58.816 "name": "raid_bdev1", 00:31:58.816 "uuid": "e5061f8a-8b03-443a-aaf1-9252fe7ee9a2", 00:31:58.816 "strip_size_kb": 0, 00:31:58.816 "state": "online", 00:31:58.816 "raid_level": "raid1", 00:31:58.816 "superblock": true, 00:31:58.816 "num_base_bdevs": 2, 00:31:58.816 "num_base_bdevs_discovered": 2, 00:31:58.816 "num_base_bdevs_operational": 2, 00:31:58.816 "base_bdevs_list": [ 00:31:58.816 { 00:31:58.816 "name": "pt1", 00:31:58.816 "uuid": "00000000-0000-0000-0000-000000000001", 00:31:58.816 "is_configured": true, 00:31:58.816 "data_offset": 256, 00:31:58.816 "data_size": 7936 00:31:58.816 }, 00:31:58.816 { 00:31:58.816 "name": "pt2", 00:31:58.816 "uuid": "00000000-0000-0000-0000-000000000002", 00:31:58.816 "is_configured": true, 00:31:58.816 "data_offset": 256, 00:31:58.816 "data_size": 7936 00:31:58.816 } 00:31:58.816 ] 00:31:58.816 }' 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:58.816 20:30:53 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:59.384 [2024-10-01 20:30:54.399710] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:31:59.384 "name": "raid_bdev1", 00:31:59.384 "aliases": [ 00:31:59.384 "e5061f8a-8b03-443a-aaf1-9252fe7ee9a2" 00:31:59.384 ], 00:31:59.384 "product_name": "Raid Volume", 00:31:59.384 "block_size": 4096, 00:31:59.384 "num_blocks": 7936, 00:31:59.384 "uuid": "e5061f8a-8b03-443a-aaf1-9252fe7ee9a2", 00:31:59.384 "assigned_rate_limits": { 00:31:59.384 "rw_ios_per_sec": 0, 00:31:59.384 "rw_mbytes_per_sec": 0, 00:31:59.384 "r_mbytes_per_sec": 0, 00:31:59.384 "w_mbytes_per_sec": 0 00:31:59.384 }, 00:31:59.384 "claimed": false, 00:31:59.384 "zoned": false, 00:31:59.384 "supported_io_types": { 00:31:59.384 "read": true, 00:31:59.384 "write": true, 00:31:59.384 "unmap": false, 00:31:59.384 "flush": false, 00:31:59.384 "reset": true, 00:31:59.384 "nvme_admin": false, 00:31:59.384 "nvme_io": false, 00:31:59.384 "nvme_io_md": false, 00:31:59.384 "write_zeroes": true, 00:31:59.384 "zcopy": false, 00:31:59.384 "get_zone_info": false, 00:31:59.384 "zone_management": false, 00:31:59.384 "zone_append": false, 00:31:59.384 "compare": false, 00:31:59.384 "compare_and_write": false, 00:31:59.384 "abort": false, 00:31:59.384 "seek_hole": false, 00:31:59.384 "seek_data": false, 00:31:59.384 "copy": false, 00:31:59.384 "nvme_iov_md": false 00:31:59.384 }, 00:31:59.384 "memory_domains": [ 00:31:59.384 { 00:31:59.384 "dma_device_id": "system", 00:31:59.384 "dma_device_type": 1 00:31:59.384 }, 00:31:59.384 { 00:31:59.384 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:31:59.384 "dma_device_type": 2 00:31:59.384 }, 00:31:59.384 { 00:31:59.384 "dma_device_id": "system", 00:31:59.384 "dma_device_type": 1 00:31:59.384 }, 00:31:59.384 { 00:31:59.384 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:31:59.384 "dma_device_type": 2 00:31:59.384 } 00:31:59.384 ], 00:31:59.384 "driver_specific": { 00:31:59.384 "raid": { 00:31:59.384 "uuid": "e5061f8a-8b03-443a-aaf1-9252fe7ee9a2", 00:31:59.384 "strip_size_kb": 0, 00:31:59.384 "state": "online", 00:31:59.384 "raid_level": "raid1", 00:31:59.384 "superblock": true, 00:31:59.384 "num_base_bdevs": 2, 00:31:59.384 "num_base_bdevs_discovered": 2, 00:31:59.384 "num_base_bdevs_operational": 2, 00:31:59.384 "base_bdevs_list": [ 00:31:59.384 { 00:31:59.384 "name": "pt1", 00:31:59.384 "uuid": "00000000-0000-0000-0000-000000000001", 00:31:59.384 "is_configured": true, 00:31:59.384 "data_offset": 256, 00:31:59.384 "data_size": 7936 00:31:59.384 }, 00:31:59.384 { 00:31:59.384 "name": "pt2", 00:31:59.384 "uuid": "00000000-0000-0000-0000-000000000002", 00:31:59.384 "is_configured": true, 00:31:59.384 "data_offset": 256, 00:31:59.384 "data_size": 7936 00:31:59.384 } 00:31:59.384 ] 00:31:59.384 } 00:31:59.384 } 00:31:59.384 }' 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:31:59.384 pt2' 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:31:59.384 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:31:59.385 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:31:59.385 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:59.385 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:59.385 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:31:59.385 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:59.644 [2024-10-01 20:30:54.660021] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # '[' e5061f8a-8b03-443a-aaf1-9252fe7ee9a2 '!=' e5061f8a-8b03-443a-aaf1-9252fe7ee9a2 ']' 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:59.644 [2024-10-01 20:30:54.715718] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:31:59.644 "name": "raid_bdev1", 00:31:59.644 "uuid": "e5061f8a-8b03-443a-aaf1-9252fe7ee9a2", 00:31:59.644 "strip_size_kb": 0, 00:31:59.644 "state": "online", 00:31:59.644 "raid_level": "raid1", 00:31:59.644 "superblock": true, 00:31:59.644 "num_base_bdevs": 2, 00:31:59.644 "num_base_bdevs_discovered": 1, 00:31:59.644 "num_base_bdevs_operational": 1, 00:31:59.644 "base_bdevs_list": [ 00:31:59.644 { 00:31:59.644 "name": null, 00:31:59.644 "uuid": "00000000-0000-0000-0000-000000000000", 00:31:59.644 "is_configured": false, 00:31:59.644 "data_offset": 0, 00:31:59.644 "data_size": 7936 00:31:59.644 }, 00:31:59.644 { 00:31:59.644 "name": "pt2", 00:31:59.644 "uuid": "00000000-0000-0000-0000-000000000002", 00:31:59.644 "is_configured": true, 00:31:59.644 "data_offset": 256, 00:31:59.644 "data_size": 7936 00:31:59.644 } 00:31:59.644 ] 00:31:59.644 }' 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:31:59.644 20:30:54 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:32:00.211 [2024-10-01 20:30:55.275705] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:00.211 [2024-10-01 20:30:55.275776] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:00.211 [2024-10-01 20:30:55.275904] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:00.211 [2024-10-01 20:30:55.275973] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:00.211 [2024-10-01 20:30:55.275995] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:32:00.211 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@519 -- # i=1 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:32:00.212 [2024-10-01 20:30:55.351717] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:32:00.212 [2024-10-01 20:30:55.351881] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:00.212 [2024-10-01 20:30:55.351913] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:32:00.212 [2024-10-01 20:30:55.351934] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:00.212 [2024-10-01 20:30:55.355447] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:00.212 [2024-10-01 20:30:55.355514] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:32:00.212 [2024-10-01 20:30:55.355621] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:32:00.212 [2024-10-01 20:30:55.355704] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:32:00.212 [2024-10-01 20:30:55.355961] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:32:00.212 [2024-10-01 20:30:55.355988] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:32:00.212 pt2 00:32:00.212 [2024-10-01 20:30:55.356320] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:32:00.212 [2024-10-01 20:30:55.356536] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:32:00.212 [2024-10-01 20:30:55.356554] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:00.212 [2024-10-01 20:30:55.356793] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:00.212 "name": "raid_bdev1", 00:32:00.212 "uuid": "e5061f8a-8b03-443a-aaf1-9252fe7ee9a2", 00:32:00.212 "strip_size_kb": 0, 00:32:00.212 "state": "online", 00:32:00.212 "raid_level": "raid1", 00:32:00.212 "superblock": true, 00:32:00.212 "num_base_bdevs": 2, 00:32:00.212 "num_base_bdevs_discovered": 1, 00:32:00.212 "num_base_bdevs_operational": 1, 00:32:00.212 "base_bdevs_list": [ 00:32:00.212 { 00:32:00.212 "name": null, 00:32:00.212 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:00.212 "is_configured": false, 00:32:00.212 "data_offset": 256, 00:32:00.212 "data_size": 7936 00:32:00.212 }, 00:32:00.212 { 00:32:00.212 "name": "pt2", 00:32:00.212 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:00.212 "is_configured": true, 00:32:00.212 "data_offset": 256, 00:32:00.212 "data_size": 7936 00:32:00.212 } 00:32:00.212 ] 00:32:00.212 }' 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:00.212 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:32:00.779 [2024-10-01 20:30:55.903990] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:00.779 [2024-10-01 20:30:55.904037] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:00.779 [2024-10-01 20:30:55.904174] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:00.779 [2024-10-01 20:30:55.904245] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:00.779 [2024-10-01 20:30:55.904261] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:32:00.779 [2024-10-01 20:30:55.972030] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:32:00.779 [2024-10-01 20:30:55.972169] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:00.779 [2024-10-01 20:30:55.972204] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:32:00.779 [2024-10-01 20:30:55.972220] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:00.779 [2024-10-01 20:30:55.975560] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:00.779 [2024-10-01 20:30:55.975609] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:32:00.779 [2024-10-01 20:30:55.975822] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:32:00.779 [2024-10-01 20:30:55.975891] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:32:00.779 [2024-10-01 20:30:55.976076] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:32:00.779 [2024-10-01 20:30:55.976111] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:00.779 [2024-10-01 20:30:55.976152] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:32:00.779 [2024-10-01 20:30:55.976260] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:32:00.779 [2024-10-01 20:30:55.976445] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:32:00.779 [2024-10-01 20:30:55.976463] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:32:00.779 pt1 00:32:00.779 [2024-10-01 20:30:55.976820] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:32:00.779 [2024-10-01 20:30:55.977046] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:32:00.779 [2024-10-01 20:30:55.977067] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:32:00.779 [2024-10-01 20:30:55.977294] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:32:00.779 20:30:55 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:01.037 20:30:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:01.037 "name": "raid_bdev1", 00:32:01.037 "uuid": "e5061f8a-8b03-443a-aaf1-9252fe7ee9a2", 00:32:01.037 "strip_size_kb": 0, 00:32:01.037 "state": "online", 00:32:01.037 "raid_level": "raid1", 00:32:01.037 "superblock": true, 00:32:01.037 "num_base_bdevs": 2, 00:32:01.037 "num_base_bdevs_discovered": 1, 00:32:01.037 "num_base_bdevs_operational": 1, 00:32:01.037 "base_bdevs_list": [ 00:32:01.037 { 00:32:01.037 "name": null, 00:32:01.037 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:01.037 "is_configured": false, 00:32:01.037 "data_offset": 256, 00:32:01.037 "data_size": 7936 00:32:01.037 }, 00:32:01.037 { 00:32:01.037 "name": "pt2", 00:32:01.037 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:01.037 "is_configured": true, 00:32:01.037 "data_offset": 256, 00:32:01.037 "data_size": 7936 00:32:01.037 } 00:32:01.037 ] 00:32:01.037 }' 00:32:01.037 20:30:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:01.037 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:32:01.295 20:30:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:32:01.295 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:01.296 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:32:01.296 20:30:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:32:01.296 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:01.296 20:30:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:32:01.296 20:30:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:32:01.296 20:30:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:32:01.296 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:01.296 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:32:01.296 [2024-10-01 20:30:56.544624] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:01.555 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:01.555 20:30:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # '[' e5061f8a-8b03-443a-aaf1-9252fe7ee9a2 '!=' e5061f8a-8b03-443a-aaf1-9252fe7ee9a2 ']' 00:32:01.555 20:30:56 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@563 -- # killprocess 87597 00:32:01.555 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@950 -- # '[' -z 87597 ']' 00:32:01.555 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@954 -- # kill -0 87597 00:32:01.555 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@955 -- # uname 00:32:01.555 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:01.555 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 87597 00:32:01.555 killing process with pid 87597 00:32:01.555 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:32:01.555 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:32:01.555 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@968 -- # echo 'killing process with pid 87597' 00:32:01.555 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@969 -- # kill 87597 00:32:01.555 [2024-10-01 20:30:56.620250] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:32:01.555 20:30:56 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@974 -- # wait 87597 00:32:01.555 [2024-10-01 20:30:56.620358] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:01.555 [2024-10-01 20:30:56.620422] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:01.555 [2024-10-01 20:30:56.620444] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:32:01.555 [2024-10-01 20:30:56.789529] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:32:03.487 20:30:58 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@565 -- # return 0 00:32:03.487 ************************************ 00:32:03.487 END TEST raid_superblock_test_4k 00:32:03.487 ************************************ 00:32:03.487 00:32:03.487 real 0m7.516s 00:32:03.487 user 0m11.319s 00:32:03.487 sys 0m1.159s 00:32:03.487 20:30:58 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1126 -- # xtrace_disable 00:32:03.487 20:30:58 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:32:03.487 20:30:58 bdev_raid -- bdev/bdev_raid.sh@999 -- # '[' true = true ']' 00:32:03.487 20:30:58 bdev_raid -- bdev/bdev_raid.sh@1000 -- # run_test raid_rebuild_test_sb_4k raid_rebuild_test raid1 2 true false true 00:32:03.487 20:30:58 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:32:03.487 20:30:58 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:32:03.487 20:30:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:32:03.487 ************************************ 00:32:03.487 START TEST raid_rebuild_test_sb_4k 00:32:03.487 ************************************ 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true false true 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # local verify=true 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # local strip_size 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@577 -- # local create_arg 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@579 -- # local data_offset 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@597 -- # raid_pid=87931 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@598 -- # waitforlisten 87931 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@831 -- # '[' -z 87931 ']' 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:03.487 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:03.487 20:30:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:03.487 [2024-10-01 20:30:58.725042] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:32:03.487 [2024-10-01 20:30:58.725522] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87931 ] 00:32:03.487 I/O size of 3145728 is greater than zero copy threshold (65536). 00:32:03.487 Zero copy mechanism will not be used. 00:32:03.747 [2024-10-01 20:30:58.904520] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:04.005 [2024-10-01 20:30:59.148557] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:32:04.264 [2024-10-01 20:30:59.353597] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:04.264 [2024-10-01 20:30:59.353989] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@864 -- # return 0 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1_malloc 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:04.833 BaseBdev1_malloc 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:04.833 [2024-10-01 20:30:59.840586] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:32:04.833 [2024-10-01 20:30:59.840762] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:04.833 [2024-10-01 20:30:59.840798] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:32:04.833 [2024-10-01 20:30:59.840839] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:04.833 [2024-10-01 20:30:59.844434] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:04.833 [2024-10-01 20:30:59.844486] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:32:04.833 BaseBdev1 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2_malloc 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:04.833 BaseBdev2_malloc 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:04.833 [2024-10-01 20:30:59.898469] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:32:04.833 [2024-10-01 20:30:59.898833] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:04.833 [2024-10-01 20:30:59.898878] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:32:04.833 [2024-10-01 20:30:59.898899] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:04.833 [2024-10-01 20:30:59.901644] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:04.833 [2024-10-01 20:30:59.901692] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:32:04.833 BaseBdev2 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -b spare_malloc 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:04.833 spare_malloc 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:04.833 spare_delay 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:04.833 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:04.833 [2024-10-01 20:30:59.958837] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:32:04.833 [2024-10-01 20:30:59.958929] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:04.833 [2024-10-01 20:30:59.958959] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:32:04.834 [2024-10-01 20:30:59.958979] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:04.834 [2024-10-01 20:30:59.961791] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:04.834 [2024-10-01 20:30:59.961832] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:32:04.834 spare 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:04.834 [2024-10-01 20:30:59.971114] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:32:04.834 [2024-10-01 20:30:59.973681] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:32:04.834 [2024-10-01 20:30:59.973984] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:32:04.834 [2024-10-01 20:30:59.974006] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:32:04.834 [2024-10-01 20:30:59.974329] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:32:04.834 [2024-10-01 20:30:59.974537] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:32:04.834 [2024-10-01 20:30:59.974552] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:32:04.834 [2024-10-01 20:30:59.974712] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:04.834 20:30:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:04.834 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:04.834 "name": "raid_bdev1", 00:32:04.834 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:04.834 "strip_size_kb": 0, 00:32:04.834 "state": "online", 00:32:04.834 "raid_level": "raid1", 00:32:04.834 "superblock": true, 00:32:04.834 "num_base_bdevs": 2, 00:32:04.834 "num_base_bdevs_discovered": 2, 00:32:04.834 "num_base_bdevs_operational": 2, 00:32:04.834 "base_bdevs_list": [ 00:32:04.834 { 00:32:04.834 "name": "BaseBdev1", 00:32:04.834 "uuid": "81710a6a-6c08-5fd7-93c8-ed0b437f38a3", 00:32:04.834 "is_configured": true, 00:32:04.834 "data_offset": 256, 00:32:04.834 "data_size": 7936 00:32:04.834 }, 00:32:04.834 { 00:32:04.834 "name": "BaseBdev2", 00:32:04.834 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:04.834 "is_configured": true, 00:32:04.834 "data_offset": 256, 00:32:04.834 "data_size": 7936 00:32:04.834 } 00:32:04.834 ] 00:32:04.834 }' 00:32:04.834 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:04.834 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:32:05.402 [2024-10-01 20:31:00.507417] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:32:05.402 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:32:05.403 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:32:05.403 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:32:05.403 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:32:05.403 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:32:05.403 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:05.403 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:32:05.662 [2024-10-01 20:31:00.907651] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:32:05.921 /dev/nbd0 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # local i 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # break 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:05.921 1+0 records in 00:32:05.921 1+0 records out 00:32:05.921 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000377176 s, 10.9 MB/s 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # size=4096 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # return 0 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:32:05.921 20:31:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:32:06.858 7936+0 records in 00:32:06.858 7936+0 records out 00:32:06.858 32505856 bytes (33 MB, 31 MiB) copied, 0.953682 s, 34.1 MB/s 00:32:06.858 20:31:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:32:06.858 20:31:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:32:06.858 20:31:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:32:06.858 20:31:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:32:06.858 20:31:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:32:06.858 20:31:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:06.858 20:31:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:32:07.118 [2024-10-01 20:31:02.251305] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:07.118 [2024-10-01 20:31:02.267645] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:07.118 "name": "raid_bdev1", 00:32:07.118 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:07.118 "strip_size_kb": 0, 00:32:07.118 "state": "online", 00:32:07.118 "raid_level": "raid1", 00:32:07.118 "superblock": true, 00:32:07.118 "num_base_bdevs": 2, 00:32:07.118 "num_base_bdevs_discovered": 1, 00:32:07.118 "num_base_bdevs_operational": 1, 00:32:07.118 "base_bdevs_list": [ 00:32:07.118 { 00:32:07.118 "name": null, 00:32:07.118 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:07.118 "is_configured": false, 00:32:07.118 "data_offset": 0, 00:32:07.118 "data_size": 7936 00:32:07.118 }, 00:32:07.118 { 00:32:07.118 "name": "BaseBdev2", 00:32:07.118 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:07.118 "is_configured": true, 00:32:07.118 "data_offset": 256, 00:32:07.118 "data_size": 7936 00:32:07.118 } 00:32:07.118 ] 00:32:07.118 }' 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:07.118 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:07.687 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:32:07.687 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:07.687 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:07.687 [2024-10-01 20:31:02.791908] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:07.687 [2024-10-01 20:31:02.808144] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d260 00:32:07.687 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:07.687 20:31:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@647 -- # sleep 1 00:32:07.687 [2024-10-01 20:31:02.810798] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:32:08.623 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:08.623 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:08.623 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:08.623 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:08.623 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:08.623 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:08.623 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:08.623 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:08.623 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:08.623 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:08.623 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:08.623 "name": "raid_bdev1", 00:32:08.623 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:08.623 "strip_size_kb": 0, 00:32:08.623 "state": "online", 00:32:08.623 "raid_level": "raid1", 00:32:08.623 "superblock": true, 00:32:08.623 "num_base_bdevs": 2, 00:32:08.623 "num_base_bdevs_discovered": 2, 00:32:08.623 "num_base_bdevs_operational": 2, 00:32:08.623 "process": { 00:32:08.623 "type": "rebuild", 00:32:08.623 "target": "spare", 00:32:08.623 "progress": { 00:32:08.623 "blocks": 2560, 00:32:08.623 "percent": 32 00:32:08.623 } 00:32:08.623 }, 00:32:08.623 "base_bdevs_list": [ 00:32:08.623 { 00:32:08.623 "name": "spare", 00:32:08.623 "uuid": "72e4ffd8-35ba-5358-8296-5cb78433764e", 00:32:08.623 "is_configured": true, 00:32:08.623 "data_offset": 256, 00:32:08.623 "data_size": 7936 00:32:08.623 }, 00:32:08.623 { 00:32:08.623 "name": "BaseBdev2", 00:32:08.623 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:08.623 "is_configured": true, 00:32:08.623 "data_offset": 256, 00:32:08.623 "data_size": 7936 00:32:08.623 } 00:32:08.623 ] 00:32:08.623 }' 00:32:08.623 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:08.882 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:08.882 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:08.882 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:08.882 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:32:08.882 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:08.882 20:31:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:08.882 [2024-10-01 20:31:03.976488] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:08.882 [2024-10-01 20:31:04.020270] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:32:08.882 [2024-10-01 20:31:04.020370] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:08.882 [2024-10-01 20:31:04.020397] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:08.882 [2024-10-01 20:31:04.020414] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:32:08.882 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:08.882 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:08.883 "name": "raid_bdev1", 00:32:08.883 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:08.883 "strip_size_kb": 0, 00:32:08.883 "state": "online", 00:32:08.883 "raid_level": "raid1", 00:32:08.883 "superblock": true, 00:32:08.883 "num_base_bdevs": 2, 00:32:08.883 "num_base_bdevs_discovered": 1, 00:32:08.883 "num_base_bdevs_operational": 1, 00:32:08.883 "base_bdevs_list": [ 00:32:08.883 { 00:32:08.883 "name": null, 00:32:08.883 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:08.883 "is_configured": false, 00:32:08.883 "data_offset": 0, 00:32:08.883 "data_size": 7936 00:32:08.883 }, 00:32:08.883 { 00:32:08.883 "name": "BaseBdev2", 00:32:08.883 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:08.883 "is_configured": true, 00:32:08.883 "data_offset": 256, 00:32:08.883 "data_size": 7936 00:32:08.883 } 00:32:08.883 ] 00:32:08.883 }' 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:08.883 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:09.461 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:09.462 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:09.462 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:09.462 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:09.462 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:09.462 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:09.462 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:09.462 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:09.462 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:09.462 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:09.462 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:09.462 "name": "raid_bdev1", 00:32:09.462 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:09.462 "strip_size_kb": 0, 00:32:09.462 "state": "online", 00:32:09.462 "raid_level": "raid1", 00:32:09.462 "superblock": true, 00:32:09.462 "num_base_bdevs": 2, 00:32:09.462 "num_base_bdevs_discovered": 1, 00:32:09.462 "num_base_bdevs_operational": 1, 00:32:09.462 "base_bdevs_list": [ 00:32:09.462 { 00:32:09.462 "name": null, 00:32:09.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:09.462 "is_configured": false, 00:32:09.462 "data_offset": 0, 00:32:09.462 "data_size": 7936 00:32:09.462 }, 00:32:09.462 { 00:32:09.462 "name": "BaseBdev2", 00:32:09.462 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:09.462 "is_configured": true, 00:32:09.462 "data_offset": 256, 00:32:09.462 "data_size": 7936 00:32:09.462 } 00:32:09.462 ] 00:32:09.462 }' 00:32:09.462 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:09.462 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:09.462 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:09.737 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:09.737 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:32:09.737 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:09.737 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:09.737 [2024-10-01 20:31:04.762686] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:09.737 [2024-10-01 20:31:04.777423] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d330 00:32:09.737 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:09.737 20:31:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@663 -- # sleep 1 00:32:09.737 [2024-10-01 20:31:04.780129] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:32:10.674 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:10.674 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:10.674 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:10.674 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:10.674 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:10.674 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:10.674 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:10.674 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:10.674 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:10.674 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:10.674 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:10.674 "name": "raid_bdev1", 00:32:10.674 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:10.674 "strip_size_kb": 0, 00:32:10.674 "state": "online", 00:32:10.674 "raid_level": "raid1", 00:32:10.674 "superblock": true, 00:32:10.674 "num_base_bdevs": 2, 00:32:10.674 "num_base_bdevs_discovered": 2, 00:32:10.674 "num_base_bdevs_operational": 2, 00:32:10.674 "process": { 00:32:10.674 "type": "rebuild", 00:32:10.674 "target": "spare", 00:32:10.674 "progress": { 00:32:10.674 "blocks": 2560, 00:32:10.674 "percent": 32 00:32:10.674 } 00:32:10.674 }, 00:32:10.674 "base_bdevs_list": [ 00:32:10.674 { 00:32:10.674 "name": "spare", 00:32:10.674 "uuid": "72e4ffd8-35ba-5358-8296-5cb78433764e", 00:32:10.674 "is_configured": true, 00:32:10.674 "data_offset": 256, 00:32:10.674 "data_size": 7936 00:32:10.674 }, 00:32:10.674 { 00:32:10.674 "name": "BaseBdev2", 00:32:10.674 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:10.674 "is_configured": true, 00:32:10.674 "data_offset": 256, 00:32:10.674 "data_size": 7936 00:32:10.674 } 00:32:10.674 ] 00:32:10.674 }' 00:32:10.674 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:10.674 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:10.674 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:32:10.934 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@706 -- # local timeout=804 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:10.934 "name": "raid_bdev1", 00:32:10.934 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:10.934 "strip_size_kb": 0, 00:32:10.934 "state": "online", 00:32:10.934 "raid_level": "raid1", 00:32:10.934 "superblock": true, 00:32:10.934 "num_base_bdevs": 2, 00:32:10.934 "num_base_bdevs_discovered": 2, 00:32:10.934 "num_base_bdevs_operational": 2, 00:32:10.934 "process": { 00:32:10.934 "type": "rebuild", 00:32:10.934 "target": "spare", 00:32:10.934 "progress": { 00:32:10.934 "blocks": 2816, 00:32:10.934 "percent": 35 00:32:10.934 } 00:32:10.934 }, 00:32:10.934 "base_bdevs_list": [ 00:32:10.934 { 00:32:10.934 "name": "spare", 00:32:10.934 "uuid": "72e4ffd8-35ba-5358-8296-5cb78433764e", 00:32:10.934 "is_configured": true, 00:32:10.934 "data_offset": 256, 00:32:10.934 "data_size": 7936 00:32:10.934 }, 00:32:10.934 { 00:32:10.934 "name": "BaseBdev2", 00:32:10.934 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:10.934 "is_configured": true, 00:32:10.934 "data_offset": 256, 00:32:10.934 "data_size": 7936 00:32:10.934 } 00:32:10.934 ] 00:32:10.934 }' 00:32:10.934 20:31:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:10.934 20:31:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:10.934 20:31:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:10.934 20:31:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:10.934 20:31:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:32:11.871 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:32:11.871 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:11.871 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:11.871 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:11.871 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:11.871 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:11.871 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:11.871 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:11.871 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:11.871 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:11.871 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:12.131 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:12.131 "name": "raid_bdev1", 00:32:12.131 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:12.131 "strip_size_kb": 0, 00:32:12.131 "state": "online", 00:32:12.131 "raid_level": "raid1", 00:32:12.131 "superblock": true, 00:32:12.131 "num_base_bdevs": 2, 00:32:12.131 "num_base_bdevs_discovered": 2, 00:32:12.131 "num_base_bdevs_operational": 2, 00:32:12.131 "process": { 00:32:12.131 "type": "rebuild", 00:32:12.131 "target": "spare", 00:32:12.131 "progress": { 00:32:12.131 "blocks": 5888, 00:32:12.131 "percent": 74 00:32:12.131 } 00:32:12.131 }, 00:32:12.131 "base_bdevs_list": [ 00:32:12.131 { 00:32:12.131 "name": "spare", 00:32:12.131 "uuid": "72e4ffd8-35ba-5358-8296-5cb78433764e", 00:32:12.131 "is_configured": true, 00:32:12.131 "data_offset": 256, 00:32:12.131 "data_size": 7936 00:32:12.131 }, 00:32:12.131 { 00:32:12.131 "name": "BaseBdev2", 00:32:12.131 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:12.131 "is_configured": true, 00:32:12.131 "data_offset": 256, 00:32:12.131 "data_size": 7936 00:32:12.131 } 00:32:12.131 ] 00:32:12.131 }' 00:32:12.131 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:12.131 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:12.131 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:12.131 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:12.131 20:31:07 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:32:12.700 [2024-10-01 20:31:07.902394] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:32:12.700 [2024-10-01 20:31:07.902521] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:32:12.700 [2024-10-01 20:31:07.902660] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:13.268 "name": "raid_bdev1", 00:32:13.268 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:13.268 "strip_size_kb": 0, 00:32:13.268 "state": "online", 00:32:13.268 "raid_level": "raid1", 00:32:13.268 "superblock": true, 00:32:13.268 "num_base_bdevs": 2, 00:32:13.268 "num_base_bdevs_discovered": 2, 00:32:13.268 "num_base_bdevs_operational": 2, 00:32:13.268 "base_bdevs_list": [ 00:32:13.268 { 00:32:13.268 "name": "spare", 00:32:13.268 "uuid": "72e4ffd8-35ba-5358-8296-5cb78433764e", 00:32:13.268 "is_configured": true, 00:32:13.268 "data_offset": 256, 00:32:13.268 "data_size": 7936 00:32:13.268 }, 00:32:13.268 { 00:32:13.268 "name": "BaseBdev2", 00:32:13.268 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:13.268 "is_configured": true, 00:32:13.268 "data_offset": 256, 00:32:13.268 "data_size": 7936 00:32:13.268 } 00:32:13.268 ] 00:32:13.268 }' 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@709 -- # break 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:13.268 "name": "raid_bdev1", 00:32:13.268 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:13.268 "strip_size_kb": 0, 00:32:13.268 "state": "online", 00:32:13.268 "raid_level": "raid1", 00:32:13.268 "superblock": true, 00:32:13.268 "num_base_bdevs": 2, 00:32:13.268 "num_base_bdevs_discovered": 2, 00:32:13.268 "num_base_bdevs_operational": 2, 00:32:13.268 "base_bdevs_list": [ 00:32:13.268 { 00:32:13.268 "name": "spare", 00:32:13.268 "uuid": "72e4ffd8-35ba-5358-8296-5cb78433764e", 00:32:13.268 "is_configured": true, 00:32:13.268 "data_offset": 256, 00:32:13.268 "data_size": 7936 00:32:13.268 }, 00:32:13.268 { 00:32:13.268 "name": "BaseBdev2", 00:32:13.268 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:13.268 "is_configured": true, 00:32:13.268 "data_offset": 256, 00:32:13.268 "data_size": 7936 00:32:13.268 } 00:32:13.268 ] 00:32:13.268 }' 00:32:13.268 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:13.527 "name": "raid_bdev1", 00:32:13.527 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:13.527 "strip_size_kb": 0, 00:32:13.527 "state": "online", 00:32:13.527 "raid_level": "raid1", 00:32:13.527 "superblock": true, 00:32:13.527 "num_base_bdevs": 2, 00:32:13.527 "num_base_bdevs_discovered": 2, 00:32:13.527 "num_base_bdevs_operational": 2, 00:32:13.527 "base_bdevs_list": [ 00:32:13.527 { 00:32:13.527 "name": "spare", 00:32:13.527 "uuid": "72e4ffd8-35ba-5358-8296-5cb78433764e", 00:32:13.527 "is_configured": true, 00:32:13.527 "data_offset": 256, 00:32:13.527 "data_size": 7936 00:32:13.527 }, 00:32:13.527 { 00:32:13.527 "name": "BaseBdev2", 00:32:13.527 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:13.527 "is_configured": true, 00:32:13.527 "data_offset": 256, 00:32:13.527 "data_size": 7936 00:32:13.527 } 00:32:13.527 ] 00:32:13.527 }' 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:13.527 20:31:08 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:14.093 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:32:14.093 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:14.093 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:14.093 [2024-10-01 20:31:09.111253] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:14.093 [2024-10-01 20:31:09.111327] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:14.093 [2024-10-01 20:31:09.111433] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:14.093 [2024-10-01 20:31:09.111522] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:14.093 [2024-10-01 20:31:09.111538] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:32:14.093 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:14.093 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:14.093 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # jq length 00:32:14.094 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:14.094 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:14.094 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:14.094 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:32:14.094 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:32:14.094 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:32:14.094 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:32:14.094 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:32:14.094 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:32:14.094 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:32:14.094 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:32:14.094 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:32:14.094 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:32:14.094 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:32:14.094 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:32:14.094 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:32:14.353 /dev/nbd0 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # local i 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # break 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:14.353 1+0 records in 00:32:14.353 1+0 records out 00:32:14.353 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00032682 s, 12.5 MB/s 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # size=4096 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # return 0 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:32:14.353 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:32:14.613 /dev/nbd1 00:32:14.613 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:32:14.613 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:32:14.613 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:32:14.613 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # local i 00:32:14.613 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:32:14.613 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:32:14.613 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:32:14.613 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # break 00:32:14.613 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:32:14.613 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:32:14.613 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:14.613 1+0 records in 00:32:14.613 1+0 records out 00:32:14.613 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00042734 s, 9.6 MB/s 00:32:14.613 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:14.872 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # size=4096 00:32:14.872 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:14.872 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:32:14.872 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # return 0 00:32:14.872 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:32:14.872 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:32:14.872 20:31:09 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:32:14.872 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:32:14.872 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:32:14.872 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:32:14.872 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:32:14.872 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:32:14.872 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:14.872 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:32:15.441 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:32:15.441 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:32:15.441 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:32:15.441 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:15.441 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:15.441 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:32:15.441 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:32:15.441 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:32:15.441 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:15.441 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:32:15.441 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:32:15.441 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:32:15.441 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:32:15.441 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:15.441 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:15.441 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:32:15.700 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:32:15.700 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:32:15.700 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:15.701 [2024-10-01 20:31:10.706856] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:32:15.701 [2024-10-01 20:31:10.706928] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:15.701 [2024-10-01 20:31:10.706971] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:32:15.701 [2024-10-01 20:31:10.706989] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:15.701 [2024-10-01 20:31:10.710256] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:15.701 [2024-10-01 20:31:10.710334] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:32:15.701 [2024-10-01 20:31:10.710523] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:32:15.701 [2024-10-01 20:31:10.710602] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:15.701 [2024-10-01 20:31:10.710836] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:32:15.701 spare 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:15.701 [2024-10-01 20:31:10.811037] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:32:15.701 [2024-10-01 20:31:10.811149] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:32:15.701 [2024-10-01 20:31:10.811679] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1b50 00:32:15.701 [2024-10-01 20:31:10.812018] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:32:15.701 [2024-10-01 20:31:10.812047] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:32:15.701 [2024-10-01 20:31:10.812304] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:15.701 "name": "raid_bdev1", 00:32:15.701 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:15.701 "strip_size_kb": 0, 00:32:15.701 "state": "online", 00:32:15.701 "raid_level": "raid1", 00:32:15.701 "superblock": true, 00:32:15.701 "num_base_bdevs": 2, 00:32:15.701 "num_base_bdevs_discovered": 2, 00:32:15.701 "num_base_bdevs_operational": 2, 00:32:15.701 "base_bdevs_list": [ 00:32:15.701 { 00:32:15.701 "name": "spare", 00:32:15.701 "uuid": "72e4ffd8-35ba-5358-8296-5cb78433764e", 00:32:15.701 "is_configured": true, 00:32:15.701 "data_offset": 256, 00:32:15.701 "data_size": 7936 00:32:15.701 }, 00:32:15.701 { 00:32:15.701 "name": "BaseBdev2", 00:32:15.701 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:15.701 "is_configured": true, 00:32:15.701 "data_offset": 256, 00:32:15.701 "data_size": 7936 00:32:15.701 } 00:32:15.701 ] 00:32:15.701 }' 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:15.701 20:31:10 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:16.270 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:16.270 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:16.270 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:16.270 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:16.270 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:16.270 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:16.270 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:16.270 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:16.270 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:16.270 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:16.270 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:16.270 "name": "raid_bdev1", 00:32:16.270 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:16.270 "strip_size_kb": 0, 00:32:16.270 "state": "online", 00:32:16.270 "raid_level": "raid1", 00:32:16.270 "superblock": true, 00:32:16.270 "num_base_bdevs": 2, 00:32:16.270 "num_base_bdevs_discovered": 2, 00:32:16.270 "num_base_bdevs_operational": 2, 00:32:16.270 "base_bdevs_list": [ 00:32:16.270 { 00:32:16.270 "name": "spare", 00:32:16.270 "uuid": "72e4ffd8-35ba-5358-8296-5cb78433764e", 00:32:16.270 "is_configured": true, 00:32:16.270 "data_offset": 256, 00:32:16.270 "data_size": 7936 00:32:16.270 }, 00:32:16.270 { 00:32:16.270 "name": "BaseBdev2", 00:32:16.270 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:16.270 "is_configured": true, 00:32:16.270 "data_offset": 256, 00:32:16.270 "data_size": 7936 00:32:16.270 } 00:32:16.270 ] 00:32:16.270 }' 00:32:16.270 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:16.270 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:16.270 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:16.270 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:16.531 [2024-10-01 20:31:11.579381] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:16.531 "name": "raid_bdev1", 00:32:16.531 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:16.531 "strip_size_kb": 0, 00:32:16.531 "state": "online", 00:32:16.531 "raid_level": "raid1", 00:32:16.531 "superblock": true, 00:32:16.531 "num_base_bdevs": 2, 00:32:16.531 "num_base_bdevs_discovered": 1, 00:32:16.531 "num_base_bdevs_operational": 1, 00:32:16.531 "base_bdevs_list": [ 00:32:16.531 { 00:32:16.531 "name": null, 00:32:16.531 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:16.531 "is_configured": false, 00:32:16.531 "data_offset": 0, 00:32:16.531 "data_size": 7936 00:32:16.531 }, 00:32:16.531 { 00:32:16.531 "name": "BaseBdev2", 00:32:16.531 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:16.531 "is_configured": true, 00:32:16.531 "data_offset": 256, 00:32:16.531 "data_size": 7936 00:32:16.531 } 00:32:16.531 ] 00:32:16.531 }' 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:16.531 20:31:11 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:17.099 20:31:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:32:17.099 20:31:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:17.099 20:31:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:17.099 [2024-10-01 20:31:12.127536] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:17.099 [2024-10-01 20:31:12.127895] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:32:17.099 [2024-10-01 20:31:12.127931] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:32:17.099 [2024-10-01 20:31:12.127998] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:17.099 [2024-10-01 20:31:12.142550] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1c20 00:32:17.099 20:31:12 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:17.099 20:31:12 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@757 -- # sleep 1 00:32:17.099 [2024-10-01 20:31:12.145530] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:32:18.035 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:18.035 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:18.035 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:18.035 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:18.035 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:18.035 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:18.035 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:18.035 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:18.035 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:18.035 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:18.035 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:18.035 "name": "raid_bdev1", 00:32:18.035 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:18.035 "strip_size_kb": 0, 00:32:18.035 "state": "online", 00:32:18.035 "raid_level": "raid1", 00:32:18.035 "superblock": true, 00:32:18.035 "num_base_bdevs": 2, 00:32:18.035 "num_base_bdevs_discovered": 2, 00:32:18.035 "num_base_bdevs_operational": 2, 00:32:18.035 "process": { 00:32:18.035 "type": "rebuild", 00:32:18.035 "target": "spare", 00:32:18.035 "progress": { 00:32:18.035 "blocks": 2560, 00:32:18.035 "percent": 32 00:32:18.035 } 00:32:18.035 }, 00:32:18.036 "base_bdevs_list": [ 00:32:18.036 { 00:32:18.036 "name": "spare", 00:32:18.036 "uuid": "72e4ffd8-35ba-5358-8296-5cb78433764e", 00:32:18.036 "is_configured": true, 00:32:18.036 "data_offset": 256, 00:32:18.036 "data_size": 7936 00:32:18.036 }, 00:32:18.036 { 00:32:18.036 "name": "BaseBdev2", 00:32:18.036 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:18.036 "is_configured": true, 00:32:18.036 "data_offset": 256, 00:32:18.036 "data_size": 7936 00:32:18.036 } 00:32:18.036 ] 00:32:18.036 }' 00:32:18.036 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:18.036 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:18.036 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:18.294 [2024-10-01 20:31:13.311154] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:18.294 [2024-10-01 20:31:13.354960] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:32:18.294 [2024-10-01 20:31:13.355067] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:18.294 [2024-10-01 20:31:13.355089] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:18.294 [2024-10-01 20:31:13.355104] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:18.294 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:18.294 "name": "raid_bdev1", 00:32:18.294 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:18.295 "strip_size_kb": 0, 00:32:18.295 "state": "online", 00:32:18.295 "raid_level": "raid1", 00:32:18.295 "superblock": true, 00:32:18.295 "num_base_bdevs": 2, 00:32:18.295 "num_base_bdevs_discovered": 1, 00:32:18.295 "num_base_bdevs_operational": 1, 00:32:18.295 "base_bdevs_list": [ 00:32:18.295 { 00:32:18.295 "name": null, 00:32:18.295 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:18.295 "is_configured": false, 00:32:18.295 "data_offset": 0, 00:32:18.295 "data_size": 7936 00:32:18.295 }, 00:32:18.295 { 00:32:18.295 "name": "BaseBdev2", 00:32:18.295 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:18.295 "is_configured": true, 00:32:18.295 "data_offset": 256, 00:32:18.295 "data_size": 7936 00:32:18.295 } 00:32:18.295 ] 00:32:18.295 }' 00:32:18.295 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:18.295 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:18.873 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:32:18.874 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:18.874 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:18.874 [2024-10-01 20:31:13.909500] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:32:18.874 [2024-10-01 20:31:13.909600] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:18.874 [2024-10-01 20:31:13.909635] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:32:18.874 [2024-10-01 20:31:13.909656] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:18.874 [2024-10-01 20:31:13.910369] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:18.874 [2024-10-01 20:31:13.910429] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:32:18.874 [2024-10-01 20:31:13.910560] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:32:18.874 [2024-10-01 20:31:13.910585] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:32:18.874 [2024-10-01 20:31:13.910601] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:32:18.874 [2024-10-01 20:31:13.910635] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:18.874 [2024-10-01 20:31:13.925864] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1cf0 00:32:18.874 spare 00:32:18.874 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:18.874 20:31:13 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@764 -- # sleep 1 00:32:18.874 [2024-10-01 20:31:13.928611] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:32:19.825 20:31:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:19.825 20:31:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:19.825 20:31:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:19.825 20:31:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:19.825 20:31:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:19.825 20:31:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:19.825 20:31:14 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:19.825 20:31:14 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:19.825 20:31:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:19.825 20:31:14 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:19.825 20:31:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:19.825 "name": "raid_bdev1", 00:32:19.825 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:19.825 "strip_size_kb": 0, 00:32:19.825 "state": "online", 00:32:19.825 "raid_level": "raid1", 00:32:19.825 "superblock": true, 00:32:19.825 "num_base_bdevs": 2, 00:32:19.825 "num_base_bdevs_discovered": 2, 00:32:19.825 "num_base_bdevs_operational": 2, 00:32:19.825 "process": { 00:32:19.825 "type": "rebuild", 00:32:19.825 "target": "spare", 00:32:19.825 "progress": { 00:32:19.825 "blocks": 2560, 00:32:19.825 "percent": 32 00:32:19.825 } 00:32:19.825 }, 00:32:19.825 "base_bdevs_list": [ 00:32:19.825 { 00:32:19.825 "name": "spare", 00:32:19.825 "uuid": "72e4ffd8-35ba-5358-8296-5cb78433764e", 00:32:19.825 "is_configured": true, 00:32:19.825 "data_offset": 256, 00:32:19.825 "data_size": 7936 00:32:19.825 }, 00:32:19.825 { 00:32:19.825 "name": "BaseBdev2", 00:32:19.825 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:19.825 "is_configured": true, 00:32:19.825 "data_offset": 256, 00:32:19.825 "data_size": 7936 00:32:19.825 } 00:32:19.825 ] 00:32:19.825 }' 00:32:19.825 20:31:14 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:19.825 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:19.825 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:20.084 [2024-10-01 20:31:15.113681] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:20.084 [2024-10-01 20:31:15.137480] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:32:20.084 [2024-10-01 20:31:15.137580] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:20.084 [2024-10-01 20:31:15.137609] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:20.084 [2024-10-01 20:31:15.137622] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:20.084 "name": "raid_bdev1", 00:32:20.084 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:20.084 "strip_size_kb": 0, 00:32:20.084 "state": "online", 00:32:20.084 "raid_level": "raid1", 00:32:20.084 "superblock": true, 00:32:20.084 "num_base_bdevs": 2, 00:32:20.084 "num_base_bdevs_discovered": 1, 00:32:20.084 "num_base_bdevs_operational": 1, 00:32:20.084 "base_bdevs_list": [ 00:32:20.084 { 00:32:20.084 "name": null, 00:32:20.084 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:20.084 "is_configured": false, 00:32:20.084 "data_offset": 0, 00:32:20.084 "data_size": 7936 00:32:20.084 }, 00:32:20.084 { 00:32:20.084 "name": "BaseBdev2", 00:32:20.084 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:20.084 "is_configured": true, 00:32:20.084 "data_offset": 256, 00:32:20.084 "data_size": 7936 00:32:20.084 } 00:32:20.084 ] 00:32:20.084 }' 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:20.084 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:20.653 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:20.653 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:20.653 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:20.653 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:20.653 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:20.653 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:20.653 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:20.654 "name": "raid_bdev1", 00:32:20.654 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:20.654 "strip_size_kb": 0, 00:32:20.654 "state": "online", 00:32:20.654 "raid_level": "raid1", 00:32:20.654 "superblock": true, 00:32:20.654 "num_base_bdevs": 2, 00:32:20.654 "num_base_bdevs_discovered": 1, 00:32:20.654 "num_base_bdevs_operational": 1, 00:32:20.654 "base_bdevs_list": [ 00:32:20.654 { 00:32:20.654 "name": null, 00:32:20.654 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:20.654 "is_configured": false, 00:32:20.654 "data_offset": 0, 00:32:20.654 "data_size": 7936 00:32:20.654 }, 00:32:20.654 { 00:32:20.654 "name": "BaseBdev2", 00:32:20.654 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:20.654 "is_configured": true, 00:32:20.654 "data_offset": 256, 00:32:20.654 "data_size": 7936 00:32:20.654 } 00:32:20.654 ] 00:32:20.654 }' 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:20.654 [2024-10-01 20:31:15.864066] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:32:20.654 [2024-10-01 20:31:15.864195] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:20.654 [2024-10-01 20:31:15.864254] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:32:20.654 [2024-10-01 20:31:15.864272] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:20.654 [2024-10-01 20:31:15.864932] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:20.654 [2024-10-01 20:31:15.864970] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:32:20.654 [2024-10-01 20:31:15.865090] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:32:20.654 [2024-10-01 20:31:15.865126] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:32:20.654 [2024-10-01 20:31:15.865140] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:32:20.654 [2024-10-01 20:31:15.865154] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:32:20.654 BaseBdev1 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:20.654 20:31:15 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@775 -- # sleep 1 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:22.031 "name": "raid_bdev1", 00:32:22.031 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:22.031 "strip_size_kb": 0, 00:32:22.031 "state": "online", 00:32:22.031 "raid_level": "raid1", 00:32:22.031 "superblock": true, 00:32:22.031 "num_base_bdevs": 2, 00:32:22.031 "num_base_bdevs_discovered": 1, 00:32:22.031 "num_base_bdevs_operational": 1, 00:32:22.031 "base_bdevs_list": [ 00:32:22.031 { 00:32:22.031 "name": null, 00:32:22.031 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:22.031 "is_configured": false, 00:32:22.031 "data_offset": 0, 00:32:22.031 "data_size": 7936 00:32:22.031 }, 00:32:22.031 { 00:32:22.031 "name": "BaseBdev2", 00:32:22.031 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:22.031 "is_configured": true, 00:32:22.031 "data_offset": 256, 00:32:22.031 "data_size": 7936 00:32:22.031 } 00:32:22.031 ] 00:32:22.031 }' 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:22.031 20:31:16 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:22.290 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:22.290 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:22.290 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:22.290 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:22.290 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:22.290 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:22.290 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:22.290 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:22.290 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:22.290 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:22.290 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:22.290 "name": "raid_bdev1", 00:32:22.290 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:22.290 "strip_size_kb": 0, 00:32:22.290 "state": "online", 00:32:22.290 "raid_level": "raid1", 00:32:22.290 "superblock": true, 00:32:22.290 "num_base_bdevs": 2, 00:32:22.290 "num_base_bdevs_discovered": 1, 00:32:22.290 "num_base_bdevs_operational": 1, 00:32:22.290 "base_bdevs_list": [ 00:32:22.290 { 00:32:22.290 "name": null, 00:32:22.290 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:22.290 "is_configured": false, 00:32:22.290 "data_offset": 0, 00:32:22.290 "data_size": 7936 00:32:22.290 }, 00:32:22.290 { 00:32:22.290 "name": "BaseBdev2", 00:32:22.290 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:22.290 "is_configured": true, 00:32:22.290 "data_offset": 256, 00:32:22.290 "data_size": 7936 00:32:22.290 } 00:32:22.290 ] 00:32:22.290 }' 00:32:22.290 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:22.290 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:22.290 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@650 -- # local es=0 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:22.548 [2024-10-01 20:31:17.565021] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:32:22.548 [2024-10-01 20:31:17.565368] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:32:22.548 [2024-10-01 20:31:17.565425] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:32:22.548 request: 00:32:22.548 { 00:32:22.548 "base_bdev": "BaseBdev1", 00:32:22.548 "raid_bdev": "raid_bdev1", 00:32:22.548 "method": "bdev_raid_add_base_bdev", 00:32:22.548 "req_id": 1 00:32:22.548 } 00:32:22.548 Got JSON-RPC error response 00:32:22.548 response: 00:32:22.548 { 00:32:22.548 "code": -22, 00:32:22.548 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:32:22.548 } 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@653 -- # es=1 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:32:22.548 20:31:17 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@779 -- # sleep 1 00:32:23.484 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:23.485 "name": "raid_bdev1", 00:32:23.485 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:23.485 "strip_size_kb": 0, 00:32:23.485 "state": "online", 00:32:23.485 "raid_level": "raid1", 00:32:23.485 "superblock": true, 00:32:23.485 "num_base_bdevs": 2, 00:32:23.485 "num_base_bdevs_discovered": 1, 00:32:23.485 "num_base_bdevs_operational": 1, 00:32:23.485 "base_bdevs_list": [ 00:32:23.485 { 00:32:23.485 "name": null, 00:32:23.485 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:23.485 "is_configured": false, 00:32:23.485 "data_offset": 0, 00:32:23.485 "data_size": 7936 00:32:23.485 }, 00:32:23.485 { 00:32:23.485 "name": "BaseBdev2", 00:32:23.485 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:23.485 "is_configured": true, 00:32:23.485 "data_offset": 256, 00:32:23.485 "data_size": 7936 00:32:23.485 } 00:32:23.485 ] 00:32:23.485 }' 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:23.485 20:31:18 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:24.053 "name": "raid_bdev1", 00:32:24.053 "uuid": "f51b17b6-86e2-4e3d-ae01-7156abd13633", 00:32:24.053 "strip_size_kb": 0, 00:32:24.053 "state": "online", 00:32:24.053 "raid_level": "raid1", 00:32:24.053 "superblock": true, 00:32:24.053 "num_base_bdevs": 2, 00:32:24.053 "num_base_bdevs_discovered": 1, 00:32:24.053 "num_base_bdevs_operational": 1, 00:32:24.053 "base_bdevs_list": [ 00:32:24.053 { 00:32:24.053 "name": null, 00:32:24.053 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:24.053 "is_configured": false, 00:32:24.053 "data_offset": 0, 00:32:24.053 "data_size": 7936 00:32:24.053 }, 00:32:24.053 { 00:32:24.053 "name": "BaseBdev2", 00:32:24.053 "uuid": "69561bc0-4848-598d-9d92-5f68d6a1d284", 00:32:24.053 "is_configured": true, 00:32:24.053 "data_offset": 256, 00:32:24.053 "data_size": 7936 00:32:24.053 } 00:32:24.053 ] 00:32:24.053 }' 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@784 -- # killprocess 87931 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@950 -- # '[' -z 87931 ']' 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@954 -- # kill -0 87931 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@955 -- # uname 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:24.053 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 87931 00:32:24.312 killing process with pid 87931 00:32:24.312 Received shutdown signal, test time was about 60.000000 seconds 00:32:24.312 00:32:24.312 Latency(us) 00:32:24.312 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:24.312 =================================================================================================================== 00:32:24.312 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:32:24.312 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:32:24.312 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:32:24.312 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@968 -- # echo 'killing process with pid 87931' 00:32:24.312 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@969 -- # kill 87931 00:32:24.312 [2024-10-01 20:31:19.305984] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:32:24.312 20:31:19 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@974 -- # wait 87931 00:32:24.312 [2024-10-01 20:31:19.306197] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:24.312 [2024-10-01 20:31:19.306261] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:24.312 [2024-10-01 20:31:19.306296] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:32:24.569 [2024-10-01 20:31:19.600729] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:32:26.472 20:31:21 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@786 -- # return 0 00:32:26.472 00:32:26.472 real 0m23.053s 00:32:26.472 user 0m30.563s 00:32:26.472 sys 0m2.914s 00:32:26.472 20:31:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1126 -- # xtrace_disable 00:32:26.472 20:31:21 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:32:26.472 ************************************ 00:32:26.472 END TEST raid_rebuild_test_sb_4k 00:32:26.472 ************************************ 00:32:26.472 20:31:21 bdev_raid -- bdev/bdev_raid.sh@1003 -- # base_malloc_params='-m 32' 00:32:26.472 20:31:21 bdev_raid -- bdev/bdev_raid.sh@1004 -- # run_test raid_state_function_test_sb_md_separate raid_state_function_test raid1 2 true 00:32:26.472 20:31:21 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:32:26.472 20:31:21 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:32:26.472 20:31:21 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:32:26.472 ************************************ 00:32:26.472 START TEST raid_state_function_test_sb_md_separate 00:32:26.472 ************************************ 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 true 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # local strip_size 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@229 -- # raid_pid=88650 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 88650' 00:32:26.472 Process raid pid: 88650 00:32:26.472 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@231 -- # waitforlisten 88650 00:32:26.731 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@831 -- # '[' -z 88650 ']' 00:32:26.731 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:26.731 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:26.731 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:26.731 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:26.731 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:26.731 20:31:21 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:26.731 [2024-10-01 20:31:21.821265] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:32:26.731 [2024-10-01 20:31:21.821470] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:26.990 [2024-10-01 20:31:21.993045] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:27.249 [2024-10-01 20:31:22.260125] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:32:27.249 [2024-10-01 20:31:22.494179] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:27.249 [2024-10-01 20:31:22.494237] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:27.817 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:27.817 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@864 -- # return 0 00:32:27.817 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:32:27.817 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:27.817 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:27.817 [2024-10-01 20:31:22.954235] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:32:27.818 [2024-10-01 20:31:22.954365] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:32:27.818 [2024-10-01 20:31:22.954383] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:32:27.818 [2024-10-01 20:31:22.954401] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:32:27.818 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:27.818 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:32:27.818 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:27.818 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:27.818 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:27.818 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:27.818 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:27.818 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:27.818 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:27.818 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:27.818 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:27.818 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:27.818 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:27.818 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:27.818 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:27.818 20:31:22 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:27.818 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:27.818 "name": "Existed_Raid", 00:32:27.818 "uuid": "37b9175d-7d78-4003-b8b9-ac703a0d8182", 00:32:27.818 "strip_size_kb": 0, 00:32:27.818 "state": "configuring", 00:32:27.818 "raid_level": "raid1", 00:32:27.818 "superblock": true, 00:32:27.818 "num_base_bdevs": 2, 00:32:27.818 "num_base_bdevs_discovered": 0, 00:32:27.818 "num_base_bdevs_operational": 2, 00:32:27.818 "base_bdevs_list": [ 00:32:27.818 { 00:32:27.818 "name": "BaseBdev1", 00:32:27.818 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:27.818 "is_configured": false, 00:32:27.818 "data_offset": 0, 00:32:27.818 "data_size": 0 00:32:27.818 }, 00:32:27.818 { 00:32:27.818 "name": "BaseBdev2", 00:32:27.818 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:27.818 "is_configured": false, 00:32:27.818 "data_offset": 0, 00:32:27.818 "data_size": 0 00:32:27.818 } 00:32:27.818 ] 00:32:27.818 }' 00:32:27.818 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:27.818 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:28.387 [2024-10-01 20:31:23.478184] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:32:28.387 [2024-10-01 20:31:23.478409] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:28.387 [2024-10-01 20:31:23.486224] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:32:28.387 [2024-10-01 20:31:23.486278] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:32:28.387 [2024-10-01 20:31:23.486295] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:32:28.387 [2024-10-01 20:31:23.486315] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:28.387 [2024-10-01 20:31:23.540023] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:32:28.387 BaseBdev1 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@901 -- # local i 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:32:28.387 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:28.388 [ 00:32:28.388 { 00:32:28.388 "name": "BaseBdev1", 00:32:28.388 "aliases": [ 00:32:28.388 "f188ef5b-548a-489b-9d3b-b25a0768331f" 00:32:28.388 ], 00:32:28.388 "product_name": "Malloc disk", 00:32:28.388 "block_size": 4096, 00:32:28.388 "num_blocks": 8192, 00:32:28.388 "uuid": "f188ef5b-548a-489b-9d3b-b25a0768331f", 00:32:28.388 "md_size": 32, 00:32:28.388 "md_interleave": false, 00:32:28.388 "dif_type": 0, 00:32:28.388 "assigned_rate_limits": { 00:32:28.388 "rw_ios_per_sec": 0, 00:32:28.388 "rw_mbytes_per_sec": 0, 00:32:28.388 "r_mbytes_per_sec": 0, 00:32:28.388 "w_mbytes_per_sec": 0 00:32:28.388 }, 00:32:28.388 "claimed": true, 00:32:28.388 "claim_type": "exclusive_write", 00:32:28.388 "zoned": false, 00:32:28.388 "supported_io_types": { 00:32:28.388 "read": true, 00:32:28.388 "write": true, 00:32:28.388 "unmap": true, 00:32:28.388 "flush": true, 00:32:28.388 "reset": true, 00:32:28.388 "nvme_admin": false, 00:32:28.388 "nvme_io": false, 00:32:28.388 "nvme_io_md": false, 00:32:28.388 "write_zeroes": true, 00:32:28.388 "zcopy": true, 00:32:28.388 "get_zone_info": false, 00:32:28.388 "zone_management": false, 00:32:28.388 "zone_append": false, 00:32:28.388 "compare": false, 00:32:28.388 "compare_and_write": false, 00:32:28.388 "abort": true, 00:32:28.388 "seek_hole": false, 00:32:28.388 "seek_data": false, 00:32:28.388 "copy": true, 00:32:28.388 "nvme_iov_md": false 00:32:28.388 }, 00:32:28.388 "memory_domains": [ 00:32:28.388 { 00:32:28.388 "dma_device_id": "system", 00:32:28.388 "dma_device_type": 1 00:32:28.388 }, 00:32:28.388 { 00:32:28.388 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:28.388 "dma_device_type": 2 00:32:28.388 } 00:32:28.388 ], 00:32:28.388 "driver_specific": {} 00:32:28.388 } 00:32:28.388 ] 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@907 -- # return 0 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:28.388 "name": "Existed_Raid", 00:32:28.388 "uuid": "925b10c0-c662-4fdf-8240-74578b1d8f7c", 00:32:28.388 "strip_size_kb": 0, 00:32:28.388 "state": "configuring", 00:32:28.388 "raid_level": "raid1", 00:32:28.388 "superblock": true, 00:32:28.388 "num_base_bdevs": 2, 00:32:28.388 "num_base_bdevs_discovered": 1, 00:32:28.388 "num_base_bdevs_operational": 2, 00:32:28.388 "base_bdevs_list": [ 00:32:28.388 { 00:32:28.388 "name": "BaseBdev1", 00:32:28.388 "uuid": "f188ef5b-548a-489b-9d3b-b25a0768331f", 00:32:28.388 "is_configured": true, 00:32:28.388 "data_offset": 256, 00:32:28.388 "data_size": 7936 00:32:28.388 }, 00:32:28.388 { 00:32:28.388 "name": "BaseBdev2", 00:32:28.388 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:28.388 "is_configured": false, 00:32:28.388 "data_offset": 0, 00:32:28.388 "data_size": 0 00:32:28.388 } 00:32:28.388 ] 00:32:28.388 }' 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:28.388 20:31:23 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:28.957 [2024-10-01 20:31:24.080316] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:32:28.957 [2024-10-01 20:31:24.080384] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:28.957 [2024-10-01 20:31:24.088420] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:32:28.957 [2024-10-01 20:31:24.091191] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:32:28.957 [2024-10-01 20:31:24.091262] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:28.957 "name": "Existed_Raid", 00:32:28.957 "uuid": "2141b49e-d2f2-482e-9a30-2c900e77219d", 00:32:28.957 "strip_size_kb": 0, 00:32:28.957 "state": "configuring", 00:32:28.957 "raid_level": "raid1", 00:32:28.957 "superblock": true, 00:32:28.957 "num_base_bdevs": 2, 00:32:28.957 "num_base_bdevs_discovered": 1, 00:32:28.957 "num_base_bdevs_operational": 2, 00:32:28.957 "base_bdevs_list": [ 00:32:28.957 { 00:32:28.957 "name": "BaseBdev1", 00:32:28.957 "uuid": "f188ef5b-548a-489b-9d3b-b25a0768331f", 00:32:28.957 "is_configured": true, 00:32:28.957 "data_offset": 256, 00:32:28.957 "data_size": 7936 00:32:28.957 }, 00:32:28.957 { 00:32:28.957 "name": "BaseBdev2", 00:32:28.957 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:28.957 "is_configured": false, 00:32:28.957 "data_offset": 0, 00:32:28.957 "data_size": 0 00:32:28.957 } 00:32:28.957 ] 00:32:28.957 }' 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:28.957 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:29.526 [2024-10-01 20:31:24.659775] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:32:29.526 [2024-10-01 20:31:24.660436] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:32:29.526 [2024-10-01 20:31:24.660463] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:32:29.526 [2024-10-01 20:31:24.660570] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:32:29.526 [2024-10-01 20:31:24.660808] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:32:29.526 BaseBdev2 00:32:29.526 [2024-10-01 20:31:24.660832] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:32:29.526 [2024-10-01 20:31:24.660955] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@901 -- # local i 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:29.526 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:29.526 [ 00:32:29.526 { 00:32:29.526 "name": "BaseBdev2", 00:32:29.526 "aliases": [ 00:32:29.526 "b3e68d41-9614-4e54-8575-20f7c3d298ec" 00:32:29.526 ], 00:32:29.526 "product_name": "Malloc disk", 00:32:29.526 "block_size": 4096, 00:32:29.526 "num_blocks": 8192, 00:32:29.526 "uuid": "b3e68d41-9614-4e54-8575-20f7c3d298ec", 00:32:29.526 "md_size": 32, 00:32:29.526 "md_interleave": false, 00:32:29.526 "dif_type": 0, 00:32:29.526 "assigned_rate_limits": { 00:32:29.526 "rw_ios_per_sec": 0, 00:32:29.526 "rw_mbytes_per_sec": 0, 00:32:29.526 "r_mbytes_per_sec": 0, 00:32:29.526 "w_mbytes_per_sec": 0 00:32:29.526 }, 00:32:29.526 "claimed": true, 00:32:29.526 "claim_type": "exclusive_write", 00:32:29.526 "zoned": false, 00:32:29.526 "supported_io_types": { 00:32:29.526 "read": true, 00:32:29.526 "write": true, 00:32:29.526 "unmap": true, 00:32:29.526 "flush": true, 00:32:29.526 "reset": true, 00:32:29.526 "nvme_admin": false, 00:32:29.526 "nvme_io": false, 00:32:29.526 "nvme_io_md": false, 00:32:29.527 "write_zeroes": true, 00:32:29.527 "zcopy": true, 00:32:29.527 "get_zone_info": false, 00:32:29.527 "zone_management": false, 00:32:29.527 "zone_append": false, 00:32:29.527 "compare": false, 00:32:29.527 "compare_and_write": false, 00:32:29.527 "abort": true, 00:32:29.527 "seek_hole": false, 00:32:29.527 "seek_data": false, 00:32:29.527 "copy": true, 00:32:29.527 "nvme_iov_md": false 00:32:29.527 }, 00:32:29.527 "memory_domains": [ 00:32:29.527 { 00:32:29.527 "dma_device_id": "system", 00:32:29.527 "dma_device_type": 1 00:32:29.527 }, 00:32:29.527 { 00:32:29.527 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:29.527 "dma_device_type": 2 00:32:29.527 } 00:32:29.527 ], 00:32:29.527 "driver_specific": {} 00:32:29.527 } 00:32:29.527 ] 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@907 -- # return 0 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:29.527 "name": "Existed_Raid", 00:32:29.527 "uuid": "2141b49e-d2f2-482e-9a30-2c900e77219d", 00:32:29.527 "strip_size_kb": 0, 00:32:29.527 "state": "online", 00:32:29.527 "raid_level": "raid1", 00:32:29.527 "superblock": true, 00:32:29.527 "num_base_bdevs": 2, 00:32:29.527 "num_base_bdevs_discovered": 2, 00:32:29.527 "num_base_bdevs_operational": 2, 00:32:29.527 "base_bdevs_list": [ 00:32:29.527 { 00:32:29.527 "name": "BaseBdev1", 00:32:29.527 "uuid": "f188ef5b-548a-489b-9d3b-b25a0768331f", 00:32:29.527 "is_configured": true, 00:32:29.527 "data_offset": 256, 00:32:29.527 "data_size": 7936 00:32:29.527 }, 00:32:29.527 { 00:32:29.527 "name": "BaseBdev2", 00:32:29.527 "uuid": "b3e68d41-9614-4e54-8575-20f7c3d298ec", 00:32:29.527 "is_configured": true, 00:32:29.527 "data_offset": 256, 00:32:29.527 "data_size": 7936 00:32:29.527 } 00:32:29.527 ] 00:32:29.527 }' 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:29.527 20:31:24 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:30.102 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:32:30.102 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:32:30.102 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:32:30.102 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:32:30.102 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:32:30.102 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:32:30.102 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:32:30.102 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:30.102 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:32:30.102 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:30.102 [2024-10-01 20:31:25.232546] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:30.102 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:30.102 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:32:30.102 "name": "Existed_Raid", 00:32:30.102 "aliases": [ 00:32:30.102 "2141b49e-d2f2-482e-9a30-2c900e77219d" 00:32:30.102 ], 00:32:30.102 "product_name": "Raid Volume", 00:32:30.102 "block_size": 4096, 00:32:30.102 "num_blocks": 7936, 00:32:30.102 "uuid": "2141b49e-d2f2-482e-9a30-2c900e77219d", 00:32:30.102 "md_size": 32, 00:32:30.102 "md_interleave": false, 00:32:30.102 "dif_type": 0, 00:32:30.102 "assigned_rate_limits": { 00:32:30.102 "rw_ios_per_sec": 0, 00:32:30.103 "rw_mbytes_per_sec": 0, 00:32:30.103 "r_mbytes_per_sec": 0, 00:32:30.103 "w_mbytes_per_sec": 0 00:32:30.103 }, 00:32:30.103 "claimed": false, 00:32:30.103 "zoned": false, 00:32:30.103 "supported_io_types": { 00:32:30.103 "read": true, 00:32:30.103 "write": true, 00:32:30.103 "unmap": false, 00:32:30.103 "flush": false, 00:32:30.103 "reset": true, 00:32:30.103 "nvme_admin": false, 00:32:30.103 "nvme_io": false, 00:32:30.103 "nvme_io_md": false, 00:32:30.103 "write_zeroes": true, 00:32:30.103 "zcopy": false, 00:32:30.103 "get_zone_info": false, 00:32:30.103 "zone_management": false, 00:32:30.103 "zone_append": false, 00:32:30.103 "compare": false, 00:32:30.103 "compare_and_write": false, 00:32:30.103 "abort": false, 00:32:30.103 "seek_hole": false, 00:32:30.103 "seek_data": false, 00:32:30.103 "copy": false, 00:32:30.103 "nvme_iov_md": false 00:32:30.103 }, 00:32:30.103 "memory_domains": [ 00:32:30.103 { 00:32:30.103 "dma_device_id": "system", 00:32:30.103 "dma_device_type": 1 00:32:30.103 }, 00:32:30.103 { 00:32:30.103 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:30.103 "dma_device_type": 2 00:32:30.103 }, 00:32:30.103 { 00:32:30.103 "dma_device_id": "system", 00:32:30.103 "dma_device_type": 1 00:32:30.103 }, 00:32:30.103 { 00:32:30.103 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:30.103 "dma_device_type": 2 00:32:30.103 } 00:32:30.103 ], 00:32:30.103 "driver_specific": { 00:32:30.103 "raid": { 00:32:30.103 "uuid": "2141b49e-d2f2-482e-9a30-2c900e77219d", 00:32:30.103 "strip_size_kb": 0, 00:32:30.103 "state": "online", 00:32:30.103 "raid_level": "raid1", 00:32:30.103 "superblock": true, 00:32:30.103 "num_base_bdevs": 2, 00:32:30.103 "num_base_bdevs_discovered": 2, 00:32:30.103 "num_base_bdevs_operational": 2, 00:32:30.103 "base_bdevs_list": [ 00:32:30.103 { 00:32:30.103 "name": "BaseBdev1", 00:32:30.103 "uuid": "f188ef5b-548a-489b-9d3b-b25a0768331f", 00:32:30.103 "is_configured": true, 00:32:30.103 "data_offset": 256, 00:32:30.103 "data_size": 7936 00:32:30.103 }, 00:32:30.103 { 00:32:30.103 "name": "BaseBdev2", 00:32:30.103 "uuid": "b3e68d41-9614-4e54-8575-20f7c3d298ec", 00:32:30.103 "is_configured": true, 00:32:30.103 "data_offset": 256, 00:32:30.103 "data_size": 7936 00:32:30.103 } 00:32:30.103 ] 00:32:30.103 } 00:32:30.103 } 00:32:30.103 }' 00:32:30.103 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:32:30.103 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:32:30.103 BaseBdev2' 00:32:30.103 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:30.372 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:32:30.372 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:30.372 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:32:30.372 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:30.372 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:30.372 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:30.372 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:30.373 [2024-10-01 20:31:25.500312] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@260 -- # local expected_state 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:30.373 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:30.632 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:30.632 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:30.632 "name": "Existed_Raid", 00:32:30.632 "uuid": "2141b49e-d2f2-482e-9a30-2c900e77219d", 00:32:30.632 "strip_size_kb": 0, 00:32:30.632 "state": "online", 00:32:30.632 "raid_level": "raid1", 00:32:30.632 "superblock": true, 00:32:30.632 "num_base_bdevs": 2, 00:32:30.632 "num_base_bdevs_discovered": 1, 00:32:30.632 "num_base_bdevs_operational": 1, 00:32:30.632 "base_bdevs_list": [ 00:32:30.632 { 00:32:30.632 "name": null, 00:32:30.632 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:30.632 "is_configured": false, 00:32:30.632 "data_offset": 0, 00:32:30.632 "data_size": 7936 00:32:30.632 }, 00:32:30.632 { 00:32:30.632 "name": "BaseBdev2", 00:32:30.632 "uuid": "b3e68d41-9614-4e54-8575-20f7c3d298ec", 00:32:30.632 "is_configured": true, 00:32:30.632 "data_offset": 256, 00:32:30.632 "data_size": 7936 00:32:30.632 } 00:32:30.632 ] 00:32:30.632 }' 00:32:30.632 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:30.632 20:31:25 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:31.199 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:32:31.199 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:32:31.199 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:32:31.199 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:31.199 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:31.199 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:31.199 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:31.199 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:32:31.199 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:32:31.199 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:32:31.199 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:31.199 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:31.199 [2024-10-01 20:31:26.222439] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:32:31.199 [2024-10-01 20:31:26.224187] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:31.200 [2024-10-01 20:31:26.327126] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:31.200 [2024-10-01 20:31:26.327202] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:31.200 [2024-10-01 20:31:26.327222] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@326 -- # killprocess 88650 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@950 -- # '[' -z 88650 ']' 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@954 -- # kill -0 88650 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@955 -- # uname 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 88650 00:32:31.200 killing process with pid 88650 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@968 -- # echo 'killing process with pid 88650' 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@969 -- # kill 88650 00:32:31.200 [2024-10-01 20:31:26.427720] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:32:31.200 20:31:26 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@974 -- # wait 88650 00:32:31.200 [2024-10-01 20:31:26.445664] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:32:33.732 20:31:28 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@328 -- # return 0 00:32:33.732 00:32:33.732 real 0m6.849s 00:32:33.732 user 0m9.571s 00:32:33.732 sys 0m0.999s 00:32:33.732 ************************************ 00:32:33.732 END TEST raid_state_function_test_sb_md_separate 00:32:33.732 ************************************ 00:32:33.732 20:31:28 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1126 -- # xtrace_disable 00:32:33.732 20:31:28 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:33.732 20:31:28 bdev_raid -- bdev/bdev_raid.sh@1005 -- # run_test raid_superblock_test_md_separate raid_superblock_test raid1 2 00:32:33.733 20:31:28 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:32:33.733 20:31:28 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:32:33.733 20:31:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:32:33.733 ************************************ 00:32:33.733 START TEST raid_superblock_test_md_separate 00:32:33.733 ************************************ 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 2 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@399 -- # local strip_size 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@412 -- # raid_pid=88915 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@413 -- # waitforlisten 88915 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@831 -- # '[' -z 88915 ']' 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:33.733 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:33.733 20:31:28 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:33.733 [2024-10-01 20:31:28.741566] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:32:33.733 [2024-10-01 20:31:28.741998] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88915 ] 00:32:33.733 [2024-10-01 20:31:28.921657] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:33.990 [2024-10-01 20:31:29.200041] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:32:34.248 [2024-10-01 20:31:29.448525] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:34.248 [2024-10-01 20:31:29.448563] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@864 -- # return 0 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc1 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:34.816 malloc1 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:34.816 [2024-10-01 20:31:29.985612] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:32:34.816 [2024-10-01 20:31:29.985934] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:34.816 [2024-10-01 20:31:29.986019] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:32:34.816 [2024-10-01 20:31:29.986288] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:34.816 [2024-10-01 20:31:29.989725] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:34.816 [2024-10-01 20:31:29.989981] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:32:34.816 pt1 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc2 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:34.816 20:31:29 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:34.816 malloc2 00:32:34.816 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:34.816 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:32:34.816 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:34.816 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:34.816 [2024-10-01 20:31:30.052553] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:32:34.816 [2024-10-01 20:31:30.052637] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:34.816 [2024-10-01 20:31:30.052685] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:32:34.816 [2024-10-01 20:31:30.052716] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:34.816 [2024-10-01 20:31:30.055962] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:34.816 [2024-10-01 20:31:30.056005] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:32:34.816 pt2 00:32:34.816 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:34.816 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:32:34.817 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:32:34.817 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:32:34.817 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:34.817 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:34.817 [2024-10-01 20:31:30.064813] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:32:34.817 [2024-10-01 20:31:30.067674] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:32:34.817 [2024-10-01 20:31:30.067968] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:32:34.817 [2024-10-01 20:31:30.067989] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:32:34.817 [2024-10-01 20:31:30.068087] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:32:34.817 [2024-10-01 20:31:30.068364] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:32:34.817 [2024-10-01 20:31:30.068399] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:32:35.076 [2024-10-01 20:31:30.068577] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:35.076 "name": "raid_bdev1", 00:32:35.076 "uuid": "78b165ab-9865-46ae-a692-d9a67a4f0b39", 00:32:35.076 "strip_size_kb": 0, 00:32:35.076 "state": "online", 00:32:35.076 "raid_level": "raid1", 00:32:35.076 "superblock": true, 00:32:35.076 "num_base_bdevs": 2, 00:32:35.076 "num_base_bdevs_discovered": 2, 00:32:35.076 "num_base_bdevs_operational": 2, 00:32:35.076 "base_bdevs_list": [ 00:32:35.076 { 00:32:35.076 "name": "pt1", 00:32:35.076 "uuid": "00000000-0000-0000-0000-000000000001", 00:32:35.076 "is_configured": true, 00:32:35.076 "data_offset": 256, 00:32:35.076 "data_size": 7936 00:32:35.076 }, 00:32:35.076 { 00:32:35.076 "name": "pt2", 00:32:35.076 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:35.076 "is_configured": true, 00:32:35.076 "data_offset": 256, 00:32:35.076 "data_size": 7936 00:32:35.076 } 00:32:35.076 ] 00:32:35.076 }' 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:35.076 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:35.644 [2024-10-01 20:31:30.609350] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:32:35.644 "name": "raid_bdev1", 00:32:35.644 "aliases": [ 00:32:35.644 "78b165ab-9865-46ae-a692-d9a67a4f0b39" 00:32:35.644 ], 00:32:35.644 "product_name": "Raid Volume", 00:32:35.644 "block_size": 4096, 00:32:35.644 "num_blocks": 7936, 00:32:35.644 "uuid": "78b165ab-9865-46ae-a692-d9a67a4f0b39", 00:32:35.644 "md_size": 32, 00:32:35.644 "md_interleave": false, 00:32:35.644 "dif_type": 0, 00:32:35.644 "assigned_rate_limits": { 00:32:35.644 "rw_ios_per_sec": 0, 00:32:35.644 "rw_mbytes_per_sec": 0, 00:32:35.644 "r_mbytes_per_sec": 0, 00:32:35.644 "w_mbytes_per_sec": 0 00:32:35.644 }, 00:32:35.644 "claimed": false, 00:32:35.644 "zoned": false, 00:32:35.644 "supported_io_types": { 00:32:35.644 "read": true, 00:32:35.644 "write": true, 00:32:35.644 "unmap": false, 00:32:35.644 "flush": false, 00:32:35.644 "reset": true, 00:32:35.644 "nvme_admin": false, 00:32:35.644 "nvme_io": false, 00:32:35.644 "nvme_io_md": false, 00:32:35.644 "write_zeroes": true, 00:32:35.644 "zcopy": false, 00:32:35.644 "get_zone_info": false, 00:32:35.644 "zone_management": false, 00:32:35.644 "zone_append": false, 00:32:35.644 "compare": false, 00:32:35.644 "compare_and_write": false, 00:32:35.644 "abort": false, 00:32:35.644 "seek_hole": false, 00:32:35.644 "seek_data": false, 00:32:35.644 "copy": false, 00:32:35.644 "nvme_iov_md": false 00:32:35.644 }, 00:32:35.644 "memory_domains": [ 00:32:35.644 { 00:32:35.644 "dma_device_id": "system", 00:32:35.644 "dma_device_type": 1 00:32:35.644 }, 00:32:35.644 { 00:32:35.644 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:35.644 "dma_device_type": 2 00:32:35.644 }, 00:32:35.644 { 00:32:35.644 "dma_device_id": "system", 00:32:35.644 "dma_device_type": 1 00:32:35.644 }, 00:32:35.644 { 00:32:35.644 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:35.644 "dma_device_type": 2 00:32:35.644 } 00:32:35.644 ], 00:32:35.644 "driver_specific": { 00:32:35.644 "raid": { 00:32:35.644 "uuid": "78b165ab-9865-46ae-a692-d9a67a4f0b39", 00:32:35.644 "strip_size_kb": 0, 00:32:35.644 "state": "online", 00:32:35.644 "raid_level": "raid1", 00:32:35.644 "superblock": true, 00:32:35.644 "num_base_bdevs": 2, 00:32:35.644 "num_base_bdevs_discovered": 2, 00:32:35.644 "num_base_bdevs_operational": 2, 00:32:35.644 "base_bdevs_list": [ 00:32:35.644 { 00:32:35.644 "name": "pt1", 00:32:35.644 "uuid": "00000000-0000-0000-0000-000000000001", 00:32:35.644 "is_configured": true, 00:32:35.644 "data_offset": 256, 00:32:35.644 "data_size": 7936 00:32:35.644 }, 00:32:35.644 { 00:32:35.644 "name": "pt2", 00:32:35.644 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:35.644 "is_configured": true, 00:32:35.644 "data_offset": 256, 00:32:35.644 "data_size": 7936 00:32:35.644 } 00:32:35.644 ] 00:32:35.644 } 00:32:35.644 } 00:32:35.644 }' 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:32:35.644 pt2' 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:35.644 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:32:35.644 [2024-10-01 20:31:30.885601] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:35.903 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:35.903 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=78b165ab-9865-46ae-a692-d9a67a4f0b39 00:32:35.903 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@436 -- # '[' -z 78b165ab-9865-46ae-a692-d9a67a4f0b39 ']' 00:32:35.903 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:32:35.903 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:35.903 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:35.903 [2024-10-01 20:31:30.937171] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:35.903 [2024-10-01 20:31:30.937223] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:35.903 [2024-10-01 20:31:30.937423] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:35.904 [2024-10-01 20:31:30.937509] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:35.904 [2024-10-01 20:31:30.937529] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:32:35.904 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:35.904 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:35.904 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:35.904 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:35.904 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:32:35.904 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:35.904 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:32:35.904 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:32:35.904 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:32:35.904 20:31:30 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:32:35.904 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:35.904 20:31:30 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@650 -- # local es=0 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:35.904 [2024-10-01 20:31:31.069351] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:32:35.904 [2024-10-01 20:31:31.072434] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:32:35.904 [2024-10-01 20:31:31.072607] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:32:35.904 [2024-10-01 20:31:31.072708] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:32:35.904 [2024-10-01 20:31:31.072781] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:35.904 [2024-10-01 20:31:31.072799] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:32:35.904 request: 00:32:35.904 { 00:32:35.904 "name": "raid_bdev1", 00:32:35.904 "raid_level": "raid1", 00:32:35.904 "base_bdevs": [ 00:32:35.904 "malloc1", 00:32:35.904 "malloc2" 00:32:35.904 ], 00:32:35.904 "superblock": false, 00:32:35.904 "method": "bdev_raid_create", 00:32:35.904 "req_id": 1 00:32:35.904 } 00:32:35.904 Got JSON-RPC error response 00:32:35.904 response: 00:32:35.904 { 00:32:35.904 "code": -17, 00:32:35.904 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:32:35.904 } 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@653 -- # es=1 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:35.904 [2024-10-01 20:31:31.133366] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:32:35.904 [2024-10-01 20:31:31.133461] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:35.904 [2024-10-01 20:31:31.133488] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:32:35.904 [2024-10-01 20:31:31.133506] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:35.904 [2024-10-01 20:31:31.136612] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:35.904 [2024-10-01 20:31:31.136684] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:32:35.904 [2024-10-01 20:31:31.136786] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:32:35.904 [2024-10-01 20:31:31.136857] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:32:35.904 pt1 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:35.904 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:36.163 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:36.163 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:36.163 "name": "raid_bdev1", 00:32:36.163 "uuid": "78b165ab-9865-46ae-a692-d9a67a4f0b39", 00:32:36.163 "strip_size_kb": 0, 00:32:36.163 "state": "configuring", 00:32:36.163 "raid_level": "raid1", 00:32:36.163 "superblock": true, 00:32:36.163 "num_base_bdevs": 2, 00:32:36.163 "num_base_bdevs_discovered": 1, 00:32:36.163 "num_base_bdevs_operational": 2, 00:32:36.163 "base_bdevs_list": [ 00:32:36.163 { 00:32:36.163 "name": "pt1", 00:32:36.163 "uuid": "00000000-0000-0000-0000-000000000001", 00:32:36.163 "is_configured": true, 00:32:36.163 "data_offset": 256, 00:32:36.163 "data_size": 7936 00:32:36.163 }, 00:32:36.163 { 00:32:36.163 "name": null, 00:32:36.163 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:36.163 "is_configured": false, 00:32:36.163 "data_offset": 256, 00:32:36.163 "data_size": 7936 00:32:36.163 } 00:32:36.163 ] 00:32:36.163 }' 00:32:36.163 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:36.163 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:36.731 [2024-10-01 20:31:31.681670] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:32:36.731 [2024-10-01 20:31:31.681773] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:36.731 [2024-10-01 20:31:31.681808] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:32:36.731 [2024-10-01 20:31:31.681829] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:36.731 [2024-10-01 20:31:31.682135] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:36.731 [2024-10-01 20:31:31.682208] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:32:36.731 [2024-10-01 20:31:31.682293] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:32:36.731 [2024-10-01 20:31:31.682328] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:32:36.731 [2024-10-01 20:31:31.682494] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:32:36.731 [2024-10-01 20:31:31.682530] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:32:36.731 [2024-10-01 20:31:31.682663] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:32:36.731 [2024-10-01 20:31:31.682809] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:32:36.731 [2024-10-01 20:31:31.682823] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:32:36.731 [2024-10-01 20:31:31.682966] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:36.731 pt2 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:36.731 "name": "raid_bdev1", 00:32:36.731 "uuid": "78b165ab-9865-46ae-a692-d9a67a4f0b39", 00:32:36.731 "strip_size_kb": 0, 00:32:36.731 "state": "online", 00:32:36.731 "raid_level": "raid1", 00:32:36.731 "superblock": true, 00:32:36.731 "num_base_bdevs": 2, 00:32:36.731 "num_base_bdevs_discovered": 2, 00:32:36.731 "num_base_bdevs_operational": 2, 00:32:36.731 "base_bdevs_list": [ 00:32:36.731 { 00:32:36.731 "name": "pt1", 00:32:36.731 "uuid": "00000000-0000-0000-0000-000000000001", 00:32:36.731 "is_configured": true, 00:32:36.731 "data_offset": 256, 00:32:36.731 "data_size": 7936 00:32:36.731 }, 00:32:36.731 { 00:32:36.731 "name": "pt2", 00:32:36.731 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:36.731 "is_configured": true, 00:32:36.731 "data_offset": 256, 00:32:36.731 "data_size": 7936 00:32:36.731 } 00:32:36.731 ] 00:32:36.731 }' 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:36.731 20:31:31 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:36.990 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:32:36.990 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:32:36.990 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:32:36.990 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:32:36.990 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:32:36.990 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:32:37.249 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:32:37.249 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:37.249 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:37.249 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:32:37.249 [2024-10-01 20:31:32.250290] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:37.249 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:37.249 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:32:37.249 "name": "raid_bdev1", 00:32:37.249 "aliases": [ 00:32:37.249 "78b165ab-9865-46ae-a692-d9a67a4f0b39" 00:32:37.249 ], 00:32:37.249 "product_name": "Raid Volume", 00:32:37.249 "block_size": 4096, 00:32:37.249 "num_blocks": 7936, 00:32:37.249 "uuid": "78b165ab-9865-46ae-a692-d9a67a4f0b39", 00:32:37.249 "md_size": 32, 00:32:37.249 "md_interleave": false, 00:32:37.249 "dif_type": 0, 00:32:37.249 "assigned_rate_limits": { 00:32:37.249 "rw_ios_per_sec": 0, 00:32:37.249 "rw_mbytes_per_sec": 0, 00:32:37.249 "r_mbytes_per_sec": 0, 00:32:37.249 "w_mbytes_per_sec": 0 00:32:37.249 }, 00:32:37.249 "claimed": false, 00:32:37.249 "zoned": false, 00:32:37.249 "supported_io_types": { 00:32:37.249 "read": true, 00:32:37.249 "write": true, 00:32:37.249 "unmap": false, 00:32:37.249 "flush": false, 00:32:37.249 "reset": true, 00:32:37.249 "nvme_admin": false, 00:32:37.249 "nvme_io": false, 00:32:37.249 "nvme_io_md": false, 00:32:37.249 "write_zeroes": true, 00:32:37.249 "zcopy": false, 00:32:37.249 "get_zone_info": false, 00:32:37.249 "zone_management": false, 00:32:37.249 "zone_append": false, 00:32:37.249 "compare": false, 00:32:37.249 "compare_and_write": false, 00:32:37.249 "abort": false, 00:32:37.249 "seek_hole": false, 00:32:37.249 "seek_data": false, 00:32:37.249 "copy": false, 00:32:37.249 "nvme_iov_md": false 00:32:37.249 }, 00:32:37.249 "memory_domains": [ 00:32:37.249 { 00:32:37.249 "dma_device_id": "system", 00:32:37.249 "dma_device_type": 1 00:32:37.250 }, 00:32:37.250 { 00:32:37.250 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:37.250 "dma_device_type": 2 00:32:37.250 }, 00:32:37.250 { 00:32:37.250 "dma_device_id": "system", 00:32:37.250 "dma_device_type": 1 00:32:37.250 }, 00:32:37.250 { 00:32:37.250 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:32:37.250 "dma_device_type": 2 00:32:37.250 } 00:32:37.250 ], 00:32:37.250 "driver_specific": { 00:32:37.250 "raid": { 00:32:37.250 "uuid": "78b165ab-9865-46ae-a692-d9a67a4f0b39", 00:32:37.250 "strip_size_kb": 0, 00:32:37.250 "state": "online", 00:32:37.250 "raid_level": "raid1", 00:32:37.250 "superblock": true, 00:32:37.250 "num_base_bdevs": 2, 00:32:37.250 "num_base_bdevs_discovered": 2, 00:32:37.250 "num_base_bdevs_operational": 2, 00:32:37.250 "base_bdevs_list": [ 00:32:37.250 { 00:32:37.250 "name": "pt1", 00:32:37.250 "uuid": "00000000-0000-0000-0000-000000000001", 00:32:37.250 "is_configured": true, 00:32:37.250 "data_offset": 256, 00:32:37.250 "data_size": 7936 00:32:37.250 }, 00:32:37.250 { 00:32:37.250 "name": "pt2", 00:32:37.250 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:37.250 "is_configured": true, 00:32:37.250 "data_offset": 256, 00:32:37.250 "data_size": 7936 00:32:37.250 } 00:32:37.250 ] 00:32:37.250 } 00:32:37.250 } 00:32:37.250 }' 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:32:37.250 pt2' 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:32:37.250 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:37.509 [2024-10-01 20:31:32.522347] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # '[' 78b165ab-9865-46ae-a692-d9a67a4f0b39 '!=' 78b165ab-9865-46ae-a692-d9a67a4f0b39 ']' 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:37.509 [2024-10-01 20:31:32.574059] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:37.509 "name": "raid_bdev1", 00:32:37.509 "uuid": "78b165ab-9865-46ae-a692-d9a67a4f0b39", 00:32:37.509 "strip_size_kb": 0, 00:32:37.509 "state": "online", 00:32:37.509 "raid_level": "raid1", 00:32:37.509 "superblock": true, 00:32:37.509 "num_base_bdevs": 2, 00:32:37.509 "num_base_bdevs_discovered": 1, 00:32:37.509 "num_base_bdevs_operational": 1, 00:32:37.509 "base_bdevs_list": [ 00:32:37.509 { 00:32:37.509 "name": null, 00:32:37.509 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:37.509 "is_configured": false, 00:32:37.509 "data_offset": 0, 00:32:37.509 "data_size": 7936 00:32:37.509 }, 00:32:37.509 { 00:32:37.509 "name": "pt2", 00:32:37.509 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:37.509 "is_configured": true, 00:32:37.509 "data_offset": 256, 00:32:37.509 "data_size": 7936 00:32:37.509 } 00:32:37.509 ] 00:32:37.509 }' 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:37.509 20:31:32 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:38.077 [2024-10-01 20:31:33.122231] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:38.077 [2024-10-01 20:31:33.122284] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:38.077 [2024-10-01 20:31:33.122405] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:38.077 [2024-10-01 20:31:33.122482] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:38.077 [2024-10-01 20:31:33.122502] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@519 -- # i=1 00:32:38.077 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:38.078 [2024-10-01 20:31:33.202193] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:32:38.078 [2024-10-01 20:31:33.202291] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:38.078 [2024-10-01 20:31:33.202331] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:32:38.078 [2024-10-01 20:31:33.202349] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:38.078 [2024-10-01 20:31:33.205694] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:38.078 [2024-10-01 20:31:33.205748] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:32:38.078 [2024-10-01 20:31:33.205821] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:32:38.078 [2024-10-01 20:31:33.205887] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:32:38.078 [2024-10-01 20:31:33.206007] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:32:38.078 [2024-10-01 20:31:33.206029] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:32:38.078 [2024-10-01 20:31:33.206115] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:32:38.078 [2024-10-01 20:31:33.206257] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:32:38.078 [2024-10-01 20:31:33.206272] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:32:38.078 pt2 00:32:38.078 [2024-10-01 20:31:33.206442] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:38.078 "name": "raid_bdev1", 00:32:38.078 "uuid": "78b165ab-9865-46ae-a692-d9a67a4f0b39", 00:32:38.078 "strip_size_kb": 0, 00:32:38.078 "state": "online", 00:32:38.078 "raid_level": "raid1", 00:32:38.078 "superblock": true, 00:32:38.078 "num_base_bdevs": 2, 00:32:38.078 "num_base_bdevs_discovered": 1, 00:32:38.078 "num_base_bdevs_operational": 1, 00:32:38.078 "base_bdevs_list": [ 00:32:38.078 { 00:32:38.078 "name": null, 00:32:38.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:38.078 "is_configured": false, 00:32:38.078 "data_offset": 256, 00:32:38.078 "data_size": 7936 00:32:38.078 }, 00:32:38.078 { 00:32:38.078 "name": "pt2", 00:32:38.078 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:38.078 "is_configured": true, 00:32:38.078 "data_offset": 256, 00:32:38.078 "data_size": 7936 00:32:38.078 } 00:32:38.078 ] 00:32:38.078 }' 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:38.078 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:38.646 [2024-10-01 20:31:33.754773] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:38.646 [2024-10-01 20:31:33.754827] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:38.646 [2024-10-01 20:31:33.754930] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:38.646 [2024-10-01 20:31:33.754999] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:38.646 [2024-10-01 20:31:33.755014] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:38.646 [2024-10-01 20:31:33.822839] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:32:38.646 [2024-10-01 20:31:33.822925] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:38.646 [2024-10-01 20:31:33.822958] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:32:38.646 [2024-10-01 20:31:33.822975] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:38.646 [2024-10-01 20:31:33.826035] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:38.646 [2024-10-01 20:31:33.826122] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:32:38.646 [2024-10-01 20:31:33.826212] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:32:38.646 [2024-10-01 20:31:33.826284] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:32:38.646 [2024-10-01 20:31:33.826482] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:32:38.646 [2024-10-01 20:31:33.826499] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:38.646 [2024-10-01 20:31:33.826529] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:32:38.646 [2024-10-01 20:31:33.826591] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:32:38.646 [2024-10-01 20:31:33.826704] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:32:38.646 [2024-10-01 20:31:33.826719] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:32:38.646 [2024-10-01 20:31:33.826847] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:32:38.646 [2024-10-01 20:31:33.826996] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:32:38.646 [2024-10-01 20:31:33.827017] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:32:38.646 [2024-10-01 20:31:33.827261] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:38.646 pt1 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:32:38.646 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:38.647 "name": "raid_bdev1", 00:32:38.647 "uuid": "78b165ab-9865-46ae-a692-d9a67a4f0b39", 00:32:38.647 "strip_size_kb": 0, 00:32:38.647 "state": "online", 00:32:38.647 "raid_level": "raid1", 00:32:38.647 "superblock": true, 00:32:38.647 "num_base_bdevs": 2, 00:32:38.647 "num_base_bdevs_discovered": 1, 00:32:38.647 "num_base_bdevs_operational": 1, 00:32:38.647 "base_bdevs_list": [ 00:32:38.647 { 00:32:38.647 "name": null, 00:32:38.647 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:38.647 "is_configured": false, 00:32:38.647 "data_offset": 256, 00:32:38.647 "data_size": 7936 00:32:38.647 }, 00:32:38.647 { 00:32:38.647 "name": "pt2", 00:32:38.647 "uuid": "00000000-0000-0000-0000-000000000002", 00:32:38.647 "is_configured": true, 00:32:38.647 "data_offset": 256, 00:32:38.647 "data_size": 7936 00:32:38.647 } 00:32:38.647 ] 00:32:38.647 }' 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:38.647 20:31:33 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:39.214 20:31:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:32:39.214 20:31:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:32:39.214 20:31:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:39.214 20:31:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:39.214 20:31:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:39.214 20:31:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:32:39.214 20:31:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:32:39.214 20:31:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:32:39.214 20:31:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:39.214 20:31:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:39.214 [2024-10-01 20:31:34.415494] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:39.214 20:31:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:39.214 20:31:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # '[' 78b165ab-9865-46ae-a692-d9a67a4f0b39 '!=' 78b165ab-9865-46ae-a692-d9a67a4f0b39 ']' 00:32:39.214 20:31:34 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@563 -- # killprocess 88915 00:32:39.214 20:31:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@950 -- # '[' -z 88915 ']' 00:32:39.214 20:31:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@954 -- # kill -0 88915 00:32:39.214 20:31:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@955 -- # uname 00:32:39.473 20:31:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:39.473 20:31:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 88915 00:32:39.473 20:31:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:32:39.473 killing process with pid 88915 00:32:39.473 20:31:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:32:39.473 20:31:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@968 -- # echo 'killing process with pid 88915' 00:32:39.474 20:31:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@969 -- # kill 88915 00:32:39.474 [2024-10-01 20:31:34.500448] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:32:39.474 20:31:34 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@974 -- # wait 88915 00:32:39.474 [2024-10-01 20:31:34.500588] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:39.474 [2024-10-01 20:31:34.500665] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:39.474 [2024-10-01 20:31:34.500689] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:32:39.474 [2024-10-01 20:31:34.715890] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:32:42.013 20:31:36 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@565 -- # return 0 00:32:42.013 00:32:42.013 real 0m8.058s 00:32:42.013 user 0m11.927s 00:32:42.013 sys 0m1.277s 00:32:42.013 20:31:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1126 -- # xtrace_disable 00:32:42.013 20:31:36 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:42.013 ************************************ 00:32:42.013 END TEST raid_superblock_test_md_separate 00:32:42.013 ************************************ 00:32:42.013 20:31:36 bdev_raid -- bdev/bdev_raid.sh@1006 -- # '[' true = true ']' 00:32:42.013 20:31:36 bdev_raid -- bdev/bdev_raid.sh@1007 -- # run_test raid_rebuild_test_sb_md_separate raid_rebuild_test raid1 2 true false true 00:32:42.013 20:31:36 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:32:42.013 20:31:36 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:32:42.013 20:31:36 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:32:42.013 ************************************ 00:32:42.013 START TEST raid_rebuild_test_sb_md_separate 00:32:42.013 ************************************ 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true false true 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # local verify=true 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # local strip_size 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@577 -- # local create_arg 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@579 -- # local data_offset 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@597 -- # raid_pid=89253 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@598 -- # waitforlisten 89253 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@831 -- # '[' -z 89253 ']' 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:42.013 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:42.013 20:31:36 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:42.013 I/O size of 3145728 is greater than zero copy threshold (65536). 00:32:42.013 Zero copy mechanism will not be used. 00:32:42.013 [2024-10-01 20:31:36.871066] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:32:42.013 [2024-10-01 20:31:36.871267] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89253 ] 00:32:42.013 [2024-10-01 20:31:37.047596] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:42.273 [2024-10-01 20:31:37.324080] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:32:42.531 [2024-10-01 20:31:37.550136] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:42.531 [2024-10-01 20:31:37.550201] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:32:42.789 20:31:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:42.789 20:31:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@864 -- # return 0 00:32:42.789 20:31:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:32:42.789 20:31:37 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1_malloc 00:32:42.789 20:31:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:42.789 20:31:37 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:43.049 BaseBdev1_malloc 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:43.049 [2024-10-01 20:31:38.054720] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:32:43.049 [2024-10-01 20:31:38.054836] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:43.049 [2024-10-01 20:31:38.054874] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:32:43.049 [2024-10-01 20:31:38.054892] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:43.049 [2024-10-01 20:31:38.057907] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:43.049 [2024-10-01 20:31:38.057950] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:32:43.049 BaseBdev1 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2_malloc 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:43.049 BaseBdev2_malloc 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:43.049 [2024-10-01 20:31:38.115171] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:32:43.049 [2024-10-01 20:31:38.115236] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:43.049 [2024-10-01 20:31:38.115264] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:32:43.049 [2024-10-01 20:31:38.115281] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:43.049 [2024-10-01 20:31:38.118310] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:43.049 [2024-10-01 20:31:38.118381] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:32:43.049 BaseBdev2 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b spare_malloc 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:43.049 spare_malloc 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:43.049 spare_delay 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:43.049 [2024-10-01 20:31:38.187234] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:32:43.049 [2024-10-01 20:31:38.187325] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:43.049 [2024-10-01 20:31:38.187365] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:32:43.049 [2024-10-01 20:31:38.187388] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:43.049 [2024-10-01 20:31:38.190273] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:43.049 [2024-10-01 20:31:38.190331] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:32:43.049 spare 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:43.049 [2024-10-01 20:31:38.195350] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:32:43.049 [2024-10-01 20:31:38.198216] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:32:43.049 [2024-10-01 20:31:38.198545] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:32:43.049 [2024-10-01 20:31:38.198574] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:32:43.049 [2024-10-01 20:31:38.198684] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:32:43.049 [2024-10-01 20:31:38.198893] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:32:43.049 [2024-10-01 20:31:38.198910] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:32:43.049 [2024-10-01 20:31:38.199047] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:43.049 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:43.049 "name": "raid_bdev1", 00:32:43.049 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:43.049 "strip_size_kb": 0, 00:32:43.049 "state": "online", 00:32:43.049 "raid_level": "raid1", 00:32:43.049 "superblock": true, 00:32:43.049 "num_base_bdevs": 2, 00:32:43.049 "num_base_bdevs_discovered": 2, 00:32:43.049 "num_base_bdevs_operational": 2, 00:32:43.049 "base_bdevs_list": [ 00:32:43.049 { 00:32:43.049 "name": "BaseBdev1", 00:32:43.050 "uuid": "13309fa3-082f-58e5-858b-2f8ba3f75c0b", 00:32:43.050 "is_configured": true, 00:32:43.050 "data_offset": 256, 00:32:43.050 "data_size": 7936 00:32:43.050 }, 00:32:43.050 { 00:32:43.050 "name": "BaseBdev2", 00:32:43.050 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:43.050 "is_configured": true, 00:32:43.050 "data_offset": 256, 00:32:43.050 "data_size": 7936 00:32:43.050 } 00:32:43.050 ] 00:32:43.050 }' 00:32:43.050 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:43.050 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:43.615 [2024-10-01 20:31:38.707840] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:43.615 20:31:38 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:32:44.180 [2024-10-01 20:31:39.147741] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:32:44.180 /dev/nbd0 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # local i 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # break 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:44.180 1+0 records in 00:32:44.180 1+0 records out 00:32:44.180 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000312459 s, 13.1 MB/s 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # size=4096 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # return 0 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:32:44.180 20:31:39 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:32:45.115 7936+0 records in 00:32:45.115 7936+0 records out 00:32:45.115 32505856 bytes (33 MB, 31 MiB) copied, 1.04601 s, 31.1 MB/s 00:32:45.115 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:32:45.115 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:32:45.115 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:32:45.115 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:32:45.115 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:32:45.115 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:45.115 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:32:45.374 [2024-10-01 20:31:40.581680] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:45.374 [2024-10-01 20:31:40.593875] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:45.374 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:45.633 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:45.633 "name": "raid_bdev1", 00:32:45.633 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:45.633 "strip_size_kb": 0, 00:32:45.633 "state": "online", 00:32:45.633 "raid_level": "raid1", 00:32:45.633 "superblock": true, 00:32:45.633 "num_base_bdevs": 2, 00:32:45.633 "num_base_bdevs_discovered": 1, 00:32:45.633 "num_base_bdevs_operational": 1, 00:32:45.633 "base_bdevs_list": [ 00:32:45.633 { 00:32:45.633 "name": null, 00:32:45.633 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:45.633 "is_configured": false, 00:32:45.633 "data_offset": 0, 00:32:45.634 "data_size": 7936 00:32:45.634 }, 00:32:45.634 { 00:32:45.634 "name": "BaseBdev2", 00:32:45.634 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:45.634 "is_configured": true, 00:32:45.634 "data_offset": 256, 00:32:45.634 "data_size": 7936 00:32:45.634 } 00:32:45.634 ] 00:32:45.634 }' 00:32:45.634 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:45.634 20:31:40 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:46.201 20:31:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:32:46.201 20:31:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:46.201 20:31:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:46.201 [2024-10-01 20:31:41.194095] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:46.201 [2024-10-01 20:31:41.209499] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d260 00:32:46.201 20:31:41 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:46.201 20:31:41 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@647 -- # sleep 1 00:32:46.201 [2024-10-01 20:31:41.212445] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:47.140 "name": "raid_bdev1", 00:32:47.140 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:47.140 "strip_size_kb": 0, 00:32:47.140 "state": "online", 00:32:47.140 "raid_level": "raid1", 00:32:47.140 "superblock": true, 00:32:47.140 "num_base_bdevs": 2, 00:32:47.140 "num_base_bdevs_discovered": 2, 00:32:47.140 "num_base_bdevs_operational": 2, 00:32:47.140 "process": { 00:32:47.140 "type": "rebuild", 00:32:47.140 "target": "spare", 00:32:47.140 "progress": { 00:32:47.140 "blocks": 2560, 00:32:47.140 "percent": 32 00:32:47.140 } 00:32:47.140 }, 00:32:47.140 "base_bdevs_list": [ 00:32:47.140 { 00:32:47.140 "name": "spare", 00:32:47.140 "uuid": "90796ee5-9388-5b9c-b3ee-75100f5eeb95", 00:32:47.140 "is_configured": true, 00:32:47.140 "data_offset": 256, 00:32:47.140 "data_size": 7936 00:32:47.140 }, 00:32:47.140 { 00:32:47.140 "name": "BaseBdev2", 00:32:47.140 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:47.140 "is_configured": true, 00:32:47.140 "data_offset": 256, 00:32:47.140 "data_size": 7936 00:32:47.140 } 00:32:47.140 ] 00:32:47.140 }' 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:47.140 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:47.140 [2024-10-01 20:31:42.386530] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:47.401 [2024-10-01 20:31:42.422512] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:32:47.401 [2024-10-01 20:31:42.422626] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:47.401 [2024-10-01 20:31:42.422677] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:47.401 [2024-10-01 20:31:42.422713] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:47.401 "name": "raid_bdev1", 00:32:47.401 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:47.401 "strip_size_kb": 0, 00:32:47.401 "state": "online", 00:32:47.401 "raid_level": "raid1", 00:32:47.401 "superblock": true, 00:32:47.401 "num_base_bdevs": 2, 00:32:47.401 "num_base_bdevs_discovered": 1, 00:32:47.401 "num_base_bdevs_operational": 1, 00:32:47.401 "base_bdevs_list": [ 00:32:47.401 { 00:32:47.401 "name": null, 00:32:47.401 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:47.401 "is_configured": false, 00:32:47.401 "data_offset": 0, 00:32:47.401 "data_size": 7936 00:32:47.401 }, 00:32:47.401 { 00:32:47.401 "name": "BaseBdev2", 00:32:47.401 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:47.401 "is_configured": true, 00:32:47.401 "data_offset": 256, 00:32:47.401 "data_size": 7936 00:32:47.401 } 00:32:47.401 ] 00:32:47.401 }' 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:47.401 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:47.971 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:47.971 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:47.971 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:47.971 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:47.971 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:47.971 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:47.971 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:47.971 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:47.971 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:47.971 20:31:42 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:47.971 20:31:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:47.971 "name": "raid_bdev1", 00:32:47.971 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:47.971 "strip_size_kb": 0, 00:32:47.971 "state": "online", 00:32:47.971 "raid_level": "raid1", 00:32:47.971 "superblock": true, 00:32:47.971 "num_base_bdevs": 2, 00:32:47.971 "num_base_bdevs_discovered": 1, 00:32:47.971 "num_base_bdevs_operational": 1, 00:32:47.971 "base_bdevs_list": [ 00:32:47.971 { 00:32:47.971 "name": null, 00:32:47.971 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:47.971 "is_configured": false, 00:32:47.971 "data_offset": 0, 00:32:47.971 "data_size": 7936 00:32:47.971 }, 00:32:47.971 { 00:32:47.971 "name": "BaseBdev2", 00:32:47.971 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:47.971 "is_configured": true, 00:32:47.971 "data_offset": 256, 00:32:47.971 "data_size": 7936 00:32:47.971 } 00:32:47.971 ] 00:32:47.971 }' 00:32:47.971 20:31:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:47.971 20:31:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:47.971 20:31:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:47.971 20:31:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:47.971 20:31:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:32:47.971 20:31:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:47.971 20:31:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:47.971 [2024-10-01 20:31:43.123817] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:47.971 [2024-10-01 20:31:43.138104] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00018d330 00:32:47.971 20:31:43 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:47.971 20:31:43 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@663 -- # sleep 1 00:32:47.971 [2024-10-01 20:31:43.141198] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:32:48.911 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:48.911 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:48.911 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:48.911 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:48.911 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:48.911 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:48.911 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:48.911 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:48.911 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:49.171 "name": "raid_bdev1", 00:32:49.171 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:49.171 "strip_size_kb": 0, 00:32:49.171 "state": "online", 00:32:49.171 "raid_level": "raid1", 00:32:49.171 "superblock": true, 00:32:49.171 "num_base_bdevs": 2, 00:32:49.171 "num_base_bdevs_discovered": 2, 00:32:49.171 "num_base_bdevs_operational": 2, 00:32:49.171 "process": { 00:32:49.171 "type": "rebuild", 00:32:49.171 "target": "spare", 00:32:49.171 "progress": { 00:32:49.171 "blocks": 2560, 00:32:49.171 "percent": 32 00:32:49.171 } 00:32:49.171 }, 00:32:49.171 "base_bdevs_list": [ 00:32:49.171 { 00:32:49.171 "name": "spare", 00:32:49.171 "uuid": "90796ee5-9388-5b9c-b3ee-75100f5eeb95", 00:32:49.171 "is_configured": true, 00:32:49.171 "data_offset": 256, 00:32:49.171 "data_size": 7936 00:32:49.171 }, 00:32:49.171 { 00:32:49.171 "name": "BaseBdev2", 00:32:49.171 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:49.171 "is_configured": true, 00:32:49.171 "data_offset": 256, 00:32:49.171 "data_size": 7936 00:32:49.171 } 00:32:49.171 ] 00:32:49.171 }' 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:32:49.171 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@706 -- # local timeout=843 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:49.171 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:49.171 "name": "raid_bdev1", 00:32:49.171 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:49.171 "strip_size_kb": 0, 00:32:49.171 "state": "online", 00:32:49.171 "raid_level": "raid1", 00:32:49.171 "superblock": true, 00:32:49.171 "num_base_bdevs": 2, 00:32:49.171 "num_base_bdevs_discovered": 2, 00:32:49.171 "num_base_bdevs_operational": 2, 00:32:49.172 "process": { 00:32:49.172 "type": "rebuild", 00:32:49.172 "target": "spare", 00:32:49.172 "progress": { 00:32:49.172 "blocks": 2816, 00:32:49.172 "percent": 35 00:32:49.172 } 00:32:49.172 }, 00:32:49.172 "base_bdevs_list": [ 00:32:49.172 { 00:32:49.172 "name": "spare", 00:32:49.172 "uuid": "90796ee5-9388-5b9c-b3ee-75100f5eeb95", 00:32:49.172 "is_configured": true, 00:32:49.172 "data_offset": 256, 00:32:49.172 "data_size": 7936 00:32:49.172 }, 00:32:49.172 { 00:32:49.172 "name": "BaseBdev2", 00:32:49.172 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:49.172 "is_configured": true, 00:32:49.172 "data_offset": 256, 00:32:49.172 "data_size": 7936 00:32:49.172 } 00:32:49.172 ] 00:32:49.172 }' 00:32:49.172 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:49.172 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:49.172 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:49.431 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:49.431 20:31:44 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:32:50.370 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:32:50.370 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:50.370 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:50.370 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:50.370 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:50.370 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:50.370 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:50.370 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:50.370 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:50.370 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:50.370 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:50.370 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:50.370 "name": "raid_bdev1", 00:32:50.370 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:50.370 "strip_size_kb": 0, 00:32:50.370 "state": "online", 00:32:50.370 "raid_level": "raid1", 00:32:50.370 "superblock": true, 00:32:50.370 "num_base_bdevs": 2, 00:32:50.370 "num_base_bdevs_discovered": 2, 00:32:50.370 "num_base_bdevs_operational": 2, 00:32:50.370 "process": { 00:32:50.370 "type": "rebuild", 00:32:50.370 "target": "spare", 00:32:50.370 "progress": { 00:32:50.370 "blocks": 5888, 00:32:50.370 "percent": 74 00:32:50.370 } 00:32:50.370 }, 00:32:50.370 "base_bdevs_list": [ 00:32:50.370 { 00:32:50.370 "name": "spare", 00:32:50.370 "uuid": "90796ee5-9388-5b9c-b3ee-75100f5eeb95", 00:32:50.370 "is_configured": true, 00:32:50.370 "data_offset": 256, 00:32:50.370 "data_size": 7936 00:32:50.370 }, 00:32:50.370 { 00:32:50.370 "name": "BaseBdev2", 00:32:50.370 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:50.370 "is_configured": true, 00:32:50.370 "data_offset": 256, 00:32:50.370 "data_size": 7936 00:32:50.370 } 00:32:50.370 ] 00:32:50.370 }' 00:32:50.370 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:50.370 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:50.370 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:50.629 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:50.629 20:31:45 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:32:51.197 [2024-10-01 20:31:46.268601] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:32:51.197 [2024-10-01 20:31:46.268849] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:32:51.197 [2024-10-01 20:31:46.269047] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:51.457 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:32:51.457 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:51.457 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:51.457 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:51.457 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:51.457 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:51.457 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:51.457 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:51.457 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:51.457 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:51.457 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:51.457 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:51.457 "name": "raid_bdev1", 00:32:51.457 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:51.457 "strip_size_kb": 0, 00:32:51.457 "state": "online", 00:32:51.457 "raid_level": "raid1", 00:32:51.457 "superblock": true, 00:32:51.457 "num_base_bdevs": 2, 00:32:51.457 "num_base_bdevs_discovered": 2, 00:32:51.457 "num_base_bdevs_operational": 2, 00:32:51.457 "base_bdevs_list": [ 00:32:51.457 { 00:32:51.457 "name": "spare", 00:32:51.457 "uuid": "90796ee5-9388-5b9c-b3ee-75100f5eeb95", 00:32:51.457 "is_configured": true, 00:32:51.457 "data_offset": 256, 00:32:51.457 "data_size": 7936 00:32:51.457 }, 00:32:51.457 { 00:32:51.457 "name": "BaseBdev2", 00:32:51.457 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:51.457 "is_configured": true, 00:32:51.457 "data_offset": 256, 00:32:51.457 "data_size": 7936 00:32:51.457 } 00:32:51.457 ] 00:32:51.457 }' 00:32:51.457 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@709 -- # break 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:51.716 "name": "raid_bdev1", 00:32:51.716 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:51.716 "strip_size_kb": 0, 00:32:51.716 "state": "online", 00:32:51.716 "raid_level": "raid1", 00:32:51.716 "superblock": true, 00:32:51.716 "num_base_bdevs": 2, 00:32:51.716 "num_base_bdevs_discovered": 2, 00:32:51.716 "num_base_bdevs_operational": 2, 00:32:51.716 "base_bdevs_list": [ 00:32:51.716 { 00:32:51.716 "name": "spare", 00:32:51.716 "uuid": "90796ee5-9388-5b9c-b3ee-75100f5eeb95", 00:32:51.716 "is_configured": true, 00:32:51.716 "data_offset": 256, 00:32:51.716 "data_size": 7936 00:32:51.716 }, 00:32:51.716 { 00:32:51.716 "name": "BaseBdev2", 00:32:51.716 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:51.716 "is_configured": true, 00:32:51.716 "data_offset": 256, 00:32:51.716 "data_size": 7936 00:32:51.716 } 00:32:51.716 ] 00:32:51.716 }' 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:51.716 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:51.975 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:51.975 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:51.975 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:51.975 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:51.975 20:31:46 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:51.975 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:51.975 "name": "raid_bdev1", 00:32:51.975 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:51.976 "strip_size_kb": 0, 00:32:51.976 "state": "online", 00:32:51.976 "raid_level": "raid1", 00:32:51.976 "superblock": true, 00:32:51.976 "num_base_bdevs": 2, 00:32:51.976 "num_base_bdevs_discovered": 2, 00:32:51.976 "num_base_bdevs_operational": 2, 00:32:51.976 "base_bdevs_list": [ 00:32:51.976 { 00:32:51.976 "name": "spare", 00:32:51.976 "uuid": "90796ee5-9388-5b9c-b3ee-75100f5eeb95", 00:32:51.976 "is_configured": true, 00:32:51.976 "data_offset": 256, 00:32:51.976 "data_size": 7936 00:32:51.976 }, 00:32:51.976 { 00:32:51.976 "name": "BaseBdev2", 00:32:51.976 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:51.976 "is_configured": true, 00:32:51.976 "data_offset": 256, 00:32:51.976 "data_size": 7936 00:32:51.976 } 00:32:51.976 ] 00:32:51.976 }' 00:32:51.976 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:51.976 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:52.543 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:32:52.543 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:52.544 [2024-10-01 20:31:47.500441] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:32:52.544 [2024-10-01 20:31:47.500642] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:32:52.544 [2024-10-01 20:31:47.500796] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:32:52.544 [2024-10-01 20:31:47.500900] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:32:52.544 [2024-10-01 20:31:47.500919] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # jq length 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:32:52.544 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:32:52.803 /dev/nbd0 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # local i 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # break 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:52.803 1+0 records in 00:32:52.803 1+0 records out 00:32:52.803 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00050056 s, 8.2 MB/s 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # size=4096 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # return 0 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:32:52.803 20:31:47 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:32:53.062 /dev/nbd1 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # local i 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # break 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:32:53.321 1+0 records in 00:32:53.321 1+0 records out 00:32:53.321 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000540764 s, 7.6 MB/s 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # size=4096 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # return 0 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:32:53.321 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:32:53.581 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:32:53.581 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:32:53.581 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:32:53.581 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:32:53.581 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:32:53.581 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:53.581 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:32:53.841 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:32:53.841 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:32:53.841 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:32:53.841 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:53.841 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:53.841 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:32:53.841 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:32:53.841 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:32:53.841 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:32:53.841 20:31:48 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:54.113 [2024-10-01 20:31:49.323714] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:32:54.113 [2024-10-01 20:31:49.323951] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:54.113 [2024-10-01 20:31:49.324137] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:32:54.113 [2024-10-01 20:31:49.324303] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:54.113 [2024-10-01 20:31:49.327574] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:54.113 [2024-10-01 20:31:49.327843] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:32:54.113 [2024-10-01 20:31:49.328056] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:32:54.113 [2024-10-01 20:31:49.328170] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:54.113 [2024-10-01 20:31:49.328404] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:32:54.113 spare 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:54.113 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:54.397 [2024-10-01 20:31:49.428536] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:32:54.397 [2024-10-01 20:31:49.428574] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:32:54.397 [2024-10-01 20:31:49.428734] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1b50 00:32:54.397 [2024-10-01 20:31:49.428972] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:32:54.397 [2024-10-01 20:31:49.428992] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:32:54.397 [2024-10-01 20:31:49.429185] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:54.397 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:54.397 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:32:54.397 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:54.397 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:54.397 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:54.397 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:54.397 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:32:54.397 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:54.397 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:54.397 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:54.397 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:54.397 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:54.397 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:54.398 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:54.398 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:54.398 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:54.398 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:54.398 "name": "raid_bdev1", 00:32:54.398 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:54.398 "strip_size_kb": 0, 00:32:54.398 "state": "online", 00:32:54.398 "raid_level": "raid1", 00:32:54.398 "superblock": true, 00:32:54.398 "num_base_bdevs": 2, 00:32:54.398 "num_base_bdevs_discovered": 2, 00:32:54.398 "num_base_bdevs_operational": 2, 00:32:54.398 "base_bdevs_list": [ 00:32:54.398 { 00:32:54.398 "name": "spare", 00:32:54.398 "uuid": "90796ee5-9388-5b9c-b3ee-75100f5eeb95", 00:32:54.398 "is_configured": true, 00:32:54.398 "data_offset": 256, 00:32:54.398 "data_size": 7936 00:32:54.398 }, 00:32:54.398 { 00:32:54.398 "name": "BaseBdev2", 00:32:54.398 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:54.398 "is_configured": true, 00:32:54.398 "data_offset": 256, 00:32:54.398 "data_size": 7936 00:32:54.398 } 00:32:54.398 ] 00:32:54.398 }' 00:32:54.398 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:54.398 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:54.966 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:54.966 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:54.966 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:54.966 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:54.966 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:54.966 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:54.966 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:54.966 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:54.966 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:54.966 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:54.966 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:54.966 "name": "raid_bdev1", 00:32:54.966 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:54.966 "strip_size_kb": 0, 00:32:54.966 "state": "online", 00:32:54.966 "raid_level": "raid1", 00:32:54.966 "superblock": true, 00:32:54.966 "num_base_bdevs": 2, 00:32:54.966 "num_base_bdevs_discovered": 2, 00:32:54.966 "num_base_bdevs_operational": 2, 00:32:54.966 "base_bdevs_list": [ 00:32:54.966 { 00:32:54.966 "name": "spare", 00:32:54.966 "uuid": "90796ee5-9388-5b9c-b3ee-75100f5eeb95", 00:32:54.966 "is_configured": true, 00:32:54.966 "data_offset": 256, 00:32:54.966 "data_size": 7936 00:32:54.966 }, 00:32:54.966 { 00:32:54.966 "name": "BaseBdev2", 00:32:54.966 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:54.966 "is_configured": true, 00:32:54.966 "data_offset": 256, 00:32:54.966 "data_size": 7936 00:32:54.966 } 00:32:54.966 ] 00:32:54.966 }' 00:32:54.966 20:31:49 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:54.966 [2024-10-01 20:31:50.152709] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:54.966 "name": "raid_bdev1", 00:32:54.966 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:54.966 "strip_size_kb": 0, 00:32:54.966 "state": "online", 00:32:54.966 "raid_level": "raid1", 00:32:54.966 "superblock": true, 00:32:54.966 "num_base_bdevs": 2, 00:32:54.966 "num_base_bdevs_discovered": 1, 00:32:54.966 "num_base_bdevs_operational": 1, 00:32:54.966 "base_bdevs_list": [ 00:32:54.966 { 00:32:54.966 "name": null, 00:32:54.966 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:54.966 "is_configured": false, 00:32:54.966 "data_offset": 0, 00:32:54.966 "data_size": 7936 00:32:54.966 }, 00:32:54.966 { 00:32:54.966 "name": "BaseBdev2", 00:32:54.966 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:54.966 "is_configured": true, 00:32:54.966 "data_offset": 256, 00:32:54.966 "data_size": 7936 00:32:54.966 } 00:32:54.966 ] 00:32:54.966 }' 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:54.966 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:55.532 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:32:55.532 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:55.532 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:55.532 [2024-10-01 20:31:50.669035] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:55.533 [2024-10-01 20:31:50.669652] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:32:55.533 [2024-10-01 20:31:50.669686] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:32:55.533 [2024-10-01 20:31:50.669757] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:55.533 [2024-10-01 20:31:50.681272] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1c20 00:32:55.533 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:55.533 20:31:50 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@757 -- # sleep 1 00:32:55.533 [2024-10-01 20:31:50.684303] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:32:56.467 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:56.467 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:56.467 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:56.467 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:56.467 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:56.467 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:56.467 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:56.467 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:56.467 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:56.467 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:56.725 "name": "raid_bdev1", 00:32:56.725 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:56.725 "strip_size_kb": 0, 00:32:56.725 "state": "online", 00:32:56.725 "raid_level": "raid1", 00:32:56.725 "superblock": true, 00:32:56.725 "num_base_bdevs": 2, 00:32:56.725 "num_base_bdevs_discovered": 2, 00:32:56.725 "num_base_bdevs_operational": 2, 00:32:56.725 "process": { 00:32:56.725 "type": "rebuild", 00:32:56.725 "target": "spare", 00:32:56.725 "progress": { 00:32:56.725 "blocks": 2560, 00:32:56.725 "percent": 32 00:32:56.725 } 00:32:56.725 }, 00:32:56.725 "base_bdevs_list": [ 00:32:56.725 { 00:32:56.725 "name": "spare", 00:32:56.725 "uuid": "90796ee5-9388-5b9c-b3ee-75100f5eeb95", 00:32:56.725 "is_configured": true, 00:32:56.725 "data_offset": 256, 00:32:56.725 "data_size": 7936 00:32:56.725 }, 00:32:56.725 { 00:32:56.725 "name": "BaseBdev2", 00:32:56.725 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:56.725 "is_configured": true, 00:32:56.725 "data_offset": 256, 00:32:56.725 "data_size": 7936 00:32:56.725 } 00:32:56.725 ] 00:32:56.725 }' 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:56.725 [2024-10-01 20:31:51.858222] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:56.725 [2024-10-01 20:31:51.895542] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:32:56.725 [2024-10-01 20:31:51.895652] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:56.725 [2024-10-01 20:31:51.895677] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:56.725 [2024-10-01 20:31:51.895699] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:56.725 "name": "raid_bdev1", 00:32:56.725 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:56.725 "strip_size_kb": 0, 00:32:56.725 "state": "online", 00:32:56.725 "raid_level": "raid1", 00:32:56.725 "superblock": true, 00:32:56.725 "num_base_bdevs": 2, 00:32:56.725 "num_base_bdevs_discovered": 1, 00:32:56.725 "num_base_bdevs_operational": 1, 00:32:56.725 "base_bdevs_list": [ 00:32:56.725 { 00:32:56.725 "name": null, 00:32:56.725 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:56.725 "is_configured": false, 00:32:56.725 "data_offset": 0, 00:32:56.725 "data_size": 7936 00:32:56.725 }, 00:32:56.725 { 00:32:56.725 "name": "BaseBdev2", 00:32:56.725 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:56.725 "is_configured": true, 00:32:56.725 "data_offset": 256, 00:32:56.725 "data_size": 7936 00:32:56.725 } 00:32:56.725 ] 00:32:56.725 }' 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:56.725 20:31:51 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:57.292 20:31:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:32:57.292 20:31:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:57.292 20:31:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:57.292 [2024-10-01 20:31:52.443062] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:32:57.292 [2024-10-01 20:31:52.443150] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:57.292 [2024-10-01 20:31:52.443187] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:32:57.292 [2024-10-01 20:31:52.443208] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:57.292 [2024-10-01 20:31:52.443555] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:57.292 [2024-10-01 20:31:52.443588] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:32:57.292 [2024-10-01 20:31:52.443673] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:32:57.292 [2024-10-01 20:31:52.443701] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:32:57.292 [2024-10-01 20:31:52.443741] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:32:57.292 [2024-10-01 20:31:52.443778] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:32:57.292 [2024-10-01 20:31:52.458655] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001c1cf0 00:32:57.292 spare 00:32:57.292 20:31:52 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:57.292 20:31:52 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@764 -- # sleep 1 00:32:57.292 [2024-10-01 20:31:52.461648] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:32:58.228 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:32:58.228 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:58.228 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:32:58.228 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:32:58.229 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:58.229 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:58.229 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:58.229 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:58.229 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:58.488 "name": "raid_bdev1", 00:32:58.488 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:58.488 "strip_size_kb": 0, 00:32:58.488 "state": "online", 00:32:58.488 "raid_level": "raid1", 00:32:58.488 "superblock": true, 00:32:58.488 "num_base_bdevs": 2, 00:32:58.488 "num_base_bdevs_discovered": 2, 00:32:58.488 "num_base_bdevs_operational": 2, 00:32:58.488 "process": { 00:32:58.488 "type": "rebuild", 00:32:58.488 "target": "spare", 00:32:58.488 "progress": { 00:32:58.488 "blocks": 2560, 00:32:58.488 "percent": 32 00:32:58.488 } 00:32:58.488 }, 00:32:58.488 "base_bdevs_list": [ 00:32:58.488 { 00:32:58.488 "name": "spare", 00:32:58.488 "uuid": "90796ee5-9388-5b9c-b3ee-75100f5eeb95", 00:32:58.488 "is_configured": true, 00:32:58.488 "data_offset": 256, 00:32:58.488 "data_size": 7936 00:32:58.488 }, 00:32:58.488 { 00:32:58.488 "name": "BaseBdev2", 00:32:58.488 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:58.488 "is_configured": true, 00:32:58.488 "data_offset": 256, 00:32:58.488 "data_size": 7936 00:32:58.488 } 00:32:58.488 ] 00:32:58.488 }' 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:58.488 [2024-10-01 20:31:53.635275] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:58.488 [2024-10-01 20:31:53.671236] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:32:58.488 [2024-10-01 20:31:53.671339] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:32:58.488 [2024-10-01 20:31:53.671368] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:32:58.488 [2024-10-01 20:31:53.671380] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:58.488 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:58.747 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:32:58.747 "name": "raid_bdev1", 00:32:58.747 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:58.747 "strip_size_kb": 0, 00:32:58.747 "state": "online", 00:32:58.747 "raid_level": "raid1", 00:32:58.747 "superblock": true, 00:32:58.747 "num_base_bdevs": 2, 00:32:58.747 "num_base_bdevs_discovered": 1, 00:32:58.747 "num_base_bdevs_operational": 1, 00:32:58.747 "base_bdevs_list": [ 00:32:58.747 { 00:32:58.747 "name": null, 00:32:58.747 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:58.747 "is_configured": false, 00:32:58.747 "data_offset": 0, 00:32:58.747 "data_size": 7936 00:32:58.747 }, 00:32:58.747 { 00:32:58.747 "name": "BaseBdev2", 00:32:58.747 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:58.747 "is_configured": true, 00:32:58.747 "data_offset": 256, 00:32:58.747 "data_size": 7936 00:32:58.747 } 00:32:58.747 ] 00:32:58.747 }' 00:32:58.747 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:32:58.747 20:31:53 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:59.006 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:32:59.006 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:32:59.006 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:32:59.006 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:32:59.006 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:32:59.006 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:32:59.006 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:32:59.006 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:59.006 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:59.006 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:59.265 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:32:59.265 "name": "raid_bdev1", 00:32:59.265 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:32:59.265 "strip_size_kb": 0, 00:32:59.265 "state": "online", 00:32:59.265 "raid_level": "raid1", 00:32:59.265 "superblock": true, 00:32:59.265 "num_base_bdevs": 2, 00:32:59.265 "num_base_bdevs_discovered": 1, 00:32:59.265 "num_base_bdevs_operational": 1, 00:32:59.265 "base_bdevs_list": [ 00:32:59.265 { 00:32:59.265 "name": null, 00:32:59.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:32:59.265 "is_configured": false, 00:32:59.265 "data_offset": 0, 00:32:59.265 "data_size": 7936 00:32:59.265 }, 00:32:59.265 { 00:32:59.265 "name": "BaseBdev2", 00:32:59.265 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:32:59.265 "is_configured": true, 00:32:59.265 "data_offset": 256, 00:32:59.265 "data_size": 7936 00:32:59.265 } 00:32:59.265 ] 00:32:59.265 }' 00:32:59.265 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:32:59.265 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:32:59.265 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:32:59.265 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:32:59.265 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:32:59.265 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:59.265 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:59.265 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:59.266 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:32:59.266 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:59.266 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:32:59.266 [2024-10-01 20:31:54.396002] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:32:59.266 [2024-10-01 20:31:54.396073] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:32:59.266 [2024-10-01 20:31:54.396123] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:32:59.266 [2024-10-01 20:31:54.396154] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:32:59.266 [2024-10-01 20:31:54.396479] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:32:59.266 [2024-10-01 20:31:54.396502] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:32:59.266 [2024-10-01 20:31:54.396588] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:32:59.266 [2024-10-01 20:31:54.396609] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:32:59.266 [2024-10-01 20:31:54.396624] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:32:59.266 [2024-10-01 20:31:54.396653] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:32:59.266 BaseBdev1 00:32:59.266 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:59.266 20:31:54 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@775 -- # sleep 1 00:33:00.202 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:33:00.202 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:00.202 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:00.202 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:00.202 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:00.202 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:33:00.202 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:00.202 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:00.202 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:00.202 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:00.202 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:00.202 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:00.202 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:00.202 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:33:00.202 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:00.461 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:00.461 "name": "raid_bdev1", 00:33:00.461 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:33:00.461 "strip_size_kb": 0, 00:33:00.461 "state": "online", 00:33:00.461 "raid_level": "raid1", 00:33:00.461 "superblock": true, 00:33:00.461 "num_base_bdevs": 2, 00:33:00.461 "num_base_bdevs_discovered": 1, 00:33:00.461 "num_base_bdevs_operational": 1, 00:33:00.461 "base_bdevs_list": [ 00:33:00.461 { 00:33:00.461 "name": null, 00:33:00.461 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:00.461 "is_configured": false, 00:33:00.461 "data_offset": 0, 00:33:00.461 "data_size": 7936 00:33:00.461 }, 00:33:00.461 { 00:33:00.461 "name": "BaseBdev2", 00:33:00.461 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:33:00.461 "is_configured": true, 00:33:00.461 "data_offset": 256, 00:33:00.461 "data_size": 7936 00:33:00.461 } 00:33:00.461 ] 00:33:00.461 }' 00:33:00.461 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:00.461 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:33:00.719 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:33:00.719 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:00.719 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:33:00.719 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:33:00.719 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:00.720 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:00.720 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:00.720 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:00.720 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:33:00.720 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:00.979 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:00.979 "name": "raid_bdev1", 00:33:00.979 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:33:00.979 "strip_size_kb": 0, 00:33:00.979 "state": "online", 00:33:00.979 "raid_level": "raid1", 00:33:00.979 "superblock": true, 00:33:00.979 "num_base_bdevs": 2, 00:33:00.979 "num_base_bdevs_discovered": 1, 00:33:00.979 "num_base_bdevs_operational": 1, 00:33:00.979 "base_bdevs_list": [ 00:33:00.979 { 00:33:00.979 "name": null, 00:33:00.979 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:00.979 "is_configured": false, 00:33:00.979 "data_offset": 0, 00:33:00.979 "data_size": 7936 00:33:00.979 }, 00:33:00.979 { 00:33:00.979 "name": "BaseBdev2", 00:33:00.979 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:33:00.979 "is_configured": true, 00:33:00.979 "data_offset": 256, 00:33:00.979 "data_size": 7936 00:33:00.979 } 00:33:00.979 ] 00:33:00.979 }' 00:33:00.979 20:31:55 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@650 -- # local es=0 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:33:00.979 [2024-10-01 20:31:56.113212] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:33:00.979 [2024-10-01 20:31:56.113439] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:33:00.979 [2024-10-01 20:31:56.113466] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:33:00.979 request: 00:33:00.979 { 00:33:00.979 "base_bdev": "BaseBdev1", 00:33:00.979 "raid_bdev": "raid_bdev1", 00:33:00.979 "method": "bdev_raid_add_base_bdev", 00:33:00.979 "req_id": 1 00:33:00.979 } 00:33:00.979 Got JSON-RPC error response 00:33:00.979 response: 00:33:00.979 { 00:33:00.979 "code": -22, 00:33:00.979 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:33:00.979 } 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@653 -- # es=1 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:33:00.979 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:33:00.980 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:33:00.980 20:31:56 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@779 -- # sleep 1 00:33:01.992 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:33:01.992 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:01.992 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:01.992 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:01.992 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:01.992 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:33:01.992 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:01.992 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:01.993 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:01.993 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:01.993 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:01.993 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:01.993 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:01.993 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:33:01.993 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:01.993 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:01.993 "name": "raid_bdev1", 00:33:01.993 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:33:01.993 "strip_size_kb": 0, 00:33:01.993 "state": "online", 00:33:01.993 "raid_level": "raid1", 00:33:01.993 "superblock": true, 00:33:01.993 "num_base_bdevs": 2, 00:33:01.993 "num_base_bdevs_discovered": 1, 00:33:01.993 "num_base_bdevs_operational": 1, 00:33:01.993 "base_bdevs_list": [ 00:33:01.993 { 00:33:01.993 "name": null, 00:33:01.993 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:01.993 "is_configured": false, 00:33:01.993 "data_offset": 0, 00:33:01.993 "data_size": 7936 00:33:01.993 }, 00:33:01.993 { 00:33:01.993 "name": "BaseBdev2", 00:33:01.993 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:33:01.993 "is_configured": true, 00:33:01.993 "data_offset": 256, 00:33:01.993 "data_size": 7936 00:33:01.993 } 00:33:01.993 ] 00:33:01.993 }' 00:33:01.993 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:01.993 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:02.559 "name": "raid_bdev1", 00:33:02.559 "uuid": "fda52b51-0036-4fba-ab3a-ad3349ea10b5", 00:33:02.559 "strip_size_kb": 0, 00:33:02.559 "state": "online", 00:33:02.559 "raid_level": "raid1", 00:33:02.559 "superblock": true, 00:33:02.559 "num_base_bdevs": 2, 00:33:02.559 "num_base_bdevs_discovered": 1, 00:33:02.559 "num_base_bdevs_operational": 1, 00:33:02.559 "base_bdevs_list": [ 00:33:02.559 { 00:33:02.559 "name": null, 00:33:02.559 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:02.559 "is_configured": false, 00:33:02.559 "data_offset": 0, 00:33:02.559 "data_size": 7936 00:33:02.559 }, 00:33:02.559 { 00:33:02.559 "name": "BaseBdev2", 00:33:02.559 "uuid": "c95fc351-486a-55dd-b8f0-912cce60bf83", 00:33:02.559 "is_configured": true, 00:33:02.559 "data_offset": 256, 00:33:02.559 "data_size": 7936 00:33:02.559 } 00:33:02.559 ] 00:33:02.559 }' 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@784 -- # killprocess 89253 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@950 -- # '[' -z 89253 ']' 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@954 -- # kill -0 89253 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@955 -- # uname 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:33:02.559 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 89253 00:33:02.817 killing process with pid 89253 00:33:02.817 Received shutdown signal, test time was about 60.000000 seconds 00:33:02.817 00:33:02.817 Latency(us) 00:33:02.817 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:02.817 =================================================================================================================== 00:33:02.817 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:33:02.817 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:33:02.817 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:33:02.817 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@968 -- # echo 'killing process with pid 89253' 00:33:02.817 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@969 -- # kill 89253 00:33:02.817 [2024-10-01 20:31:57.827002] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:33:02.817 20:31:57 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@974 -- # wait 89253 00:33:02.817 [2024-10-01 20:31:57.827201] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:02.817 [2024-10-01 20:31:57.827271] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:02.817 [2024-10-01 20:31:57.827291] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:33:03.075 [2024-10-01 20:31:58.125383] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:33:04.981 ************************************ 00:33:04.981 END TEST raid_rebuild_test_sb_md_separate 00:33:04.981 ************************************ 00:33:04.981 20:32:00 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@786 -- # return 0 00:33:04.981 00:33:04.981 real 0m23.297s 00:33:04.981 user 0m30.982s 00:33:04.981 sys 0m3.045s 00:33:04.981 20:32:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1126 -- # xtrace_disable 00:33:04.981 20:32:00 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:33:04.981 20:32:00 bdev_raid -- bdev/bdev_raid.sh@1010 -- # base_malloc_params='-m 32 -i' 00:33:04.981 20:32:00 bdev_raid -- bdev/bdev_raid.sh@1011 -- # run_test raid_state_function_test_sb_md_interleaved raid_state_function_test raid1 2 true 00:33:04.981 20:32:00 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:33:04.981 20:32:00 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:33:04.981 20:32:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:33:04.981 ************************************ 00:33:04.981 START TEST raid_state_function_test_sb_md_interleaved 00:33:04.981 ************************************ 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 true 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # local strip_size 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:33:04.981 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:33:04.982 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:33:04.982 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:33:04.982 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:33:04.982 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@229 -- # raid_pid=89976 00:33:04.982 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:33:04.982 Process raid pid: 89976 00:33:04.982 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:04.982 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 89976' 00:33:04.982 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@231 -- # waitforlisten 89976 00:33:04.982 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@831 -- # '[' -z 89976 ']' 00:33:04.982 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:04.982 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@836 -- # local max_retries=100 00:33:04.982 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:04.982 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # xtrace_disable 00:33:04.982 20:32:00 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:05.240 [2024-10-01 20:32:00.235788] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:33:05.240 [2024-10-01 20:32:00.236269] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:33:05.240 [2024-10-01 20:32:00.423763] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:05.498 [2024-10-01 20:32:00.696829] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:33:05.758 [2024-10-01 20:32:00.924128] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:33:05.758 [2024-10-01 20:32:00.924181] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # return 0 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:06.325 [2024-10-01 20:32:01.441140] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:33:06.325 [2024-10-01 20:32:01.441215] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:33:06.325 [2024-10-01 20:32:01.441236] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:33:06.325 [2024-10-01 20:32:01.441258] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:06.325 "name": "Existed_Raid", 00:33:06.325 "uuid": "0e212ed9-0f04-4a6e-a271-faf42f5fbe2d", 00:33:06.325 "strip_size_kb": 0, 00:33:06.325 "state": "configuring", 00:33:06.325 "raid_level": "raid1", 00:33:06.325 "superblock": true, 00:33:06.325 "num_base_bdevs": 2, 00:33:06.325 "num_base_bdevs_discovered": 0, 00:33:06.325 "num_base_bdevs_operational": 2, 00:33:06.325 "base_bdevs_list": [ 00:33:06.325 { 00:33:06.325 "name": "BaseBdev1", 00:33:06.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:06.325 "is_configured": false, 00:33:06.325 "data_offset": 0, 00:33:06.325 "data_size": 0 00:33:06.325 }, 00:33:06.325 { 00:33:06.325 "name": "BaseBdev2", 00:33:06.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:06.325 "is_configured": false, 00:33:06.325 "data_offset": 0, 00:33:06.325 "data_size": 0 00:33:06.325 } 00:33:06.325 ] 00:33:06.325 }' 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:06.325 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:06.892 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:33:06.892 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:06.892 20:32:01 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:06.892 [2024-10-01 20:32:02.005170] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:33:06.892 [2024-10-01 20:32:02.005375] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name Existed_Raid, state configuring 00:33:06.892 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:06.892 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:33:06.892 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:06.892 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:06.892 [2024-10-01 20:32:02.013173] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:33:06.892 [2024-10-01 20:32:02.013239] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:33:06.893 [2024-10-01 20:32:02.013259] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:33:06.893 [2024-10-01 20:32:02.013282] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:06.893 [2024-10-01 20:32:02.059826] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:33:06.893 BaseBdev1 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@901 -- # local i 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:06.893 [ 00:33:06.893 { 00:33:06.893 "name": "BaseBdev1", 00:33:06.893 "aliases": [ 00:33:06.893 "b3f6c06c-3d04-42fb-bfc9-009915584700" 00:33:06.893 ], 00:33:06.893 "product_name": "Malloc disk", 00:33:06.893 "block_size": 4128, 00:33:06.893 "num_blocks": 8192, 00:33:06.893 "uuid": "b3f6c06c-3d04-42fb-bfc9-009915584700", 00:33:06.893 "md_size": 32, 00:33:06.893 "md_interleave": true, 00:33:06.893 "dif_type": 0, 00:33:06.893 "assigned_rate_limits": { 00:33:06.893 "rw_ios_per_sec": 0, 00:33:06.893 "rw_mbytes_per_sec": 0, 00:33:06.893 "r_mbytes_per_sec": 0, 00:33:06.893 "w_mbytes_per_sec": 0 00:33:06.893 }, 00:33:06.893 "claimed": true, 00:33:06.893 "claim_type": "exclusive_write", 00:33:06.893 "zoned": false, 00:33:06.893 "supported_io_types": { 00:33:06.893 "read": true, 00:33:06.893 "write": true, 00:33:06.893 "unmap": true, 00:33:06.893 "flush": true, 00:33:06.893 "reset": true, 00:33:06.893 "nvme_admin": false, 00:33:06.893 "nvme_io": false, 00:33:06.893 "nvme_io_md": false, 00:33:06.893 "write_zeroes": true, 00:33:06.893 "zcopy": true, 00:33:06.893 "get_zone_info": false, 00:33:06.893 "zone_management": false, 00:33:06.893 "zone_append": false, 00:33:06.893 "compare": false, 00:33:06.893 "compare_and_write": false, 00:33:06.893 "abort": true, 00:33:06.893 "seek_hole": false, 00:33:06.893 "seek_data": false, 00:33:06.893 "copy": true, 00:33:06.893 "nvme_iov_md": false 00:33:06.893 }, 00:33:06.893 "memory_domains": [ 00:33:06.893 { 00:33:06.893 "dma_device_id": "system", 00:33:06.893 "dma_device_type": 1 00:33:06.893 }, 00:33:06.893 { 00:33:06.893 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:06.893 "dma_device_type": 2 00:33:06.893 } 00:33:06.893 ], 00:33:06.893 "driver_specific": {} 00:33:06.893 } 00:33:06.893 ] 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@907 -- # return 0 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:06.893 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:07.151 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:07.151 "name": "Existed_Raid", 00:33:07.151 "uuid": "976ce6be-9208-450c-8632-a7a01dd2cec0", 00:33:07.151 "strip_size_kb": 0, 00:33:07.151 "state": "configuring", 00:33:07.151 "raid_level": "raid1", 00:33:07.151 "superblock": true, 00:33:07.151 "num_base_bdevs": 2, 00:33:07.151 "num_base_bdevs_discovered": 1, 00:33:07.151 "num_base_bdevs_operational": 2, 00:33:07.151 "base_bdevs_list": [ 00:33:07.151 { 00:33:07.151 "name": "BaseBdev1", 00:33:07.151 "uuid": "b3f6c06c-3d04-42fb-bfc9-009915584700", 00:33:07.151 "is_configured": true, 00:33:07.151 "data_offset": 256, 00:33:07.151 "data_size": 7936 00:33:07.151 }, 00:33:07.151 { 00:33:07.151 "name": "BaseBdev2", 00:33:07.151 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:07.151 "is_configured": false, 00:33:07.151 "data_offset": 0, 00:33:07.151 "data_size": 0 00:33:07.151 } 00:33:07.151 ] 00:33:07.151 }' 00:33:07.151 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:07.151 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:07.409 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:33:07.409 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:07.409 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:07.409 [2024-10-01 20:32:02.648176] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:33:07.409 [2024-10-01 20:32:02.648305] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name Existed_Raid, state configuring 00:33:07.409 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:07.409 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:33:07.409 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:07.409 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:07.409 [2024-10-01 20:32:02.656156] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:33:07.409 [2024-10-01 20:32:02.659202] bdev.c:8320:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:33:07.409 [2024-10-01 20:32:02.659434] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:33:07.409 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:07.667 "name": "Existed_Raid", 00:33:07.667 "uuid": "36bcab26-fbb4-4868-b384-5c99c5b1971a", 00:33:07.667 "strip_size_kb": 0, 00:33:07.667 "state": "configuring", 00:33:07.667 "raid_level": "raid1", 00:33:07.667 "superblock": true, 00:33:07.667 "num_base_bdevs": 2, 00:33:07.667 "num_base_bdevs_discovered": 1, 00:33:07.667 "num_base_bdevs_operational": 2, 00:33:07.667 "base_bdevs_list": [ 00:33:07.667 { 00:33:07.667 "name": "BaseBdev1", 00:33:07.667 "uuid": "b3f6c06c-3d04-42fb-bfc9-009915584700", 00:33:07.667 "is_configured": true, 00:33:07.667 "data_offset": 256, 00:33:07.667 "data_size": 7936 00:33:07.667 }, 00:33:07.667 { 00:33:07.667 "name": "BaseBdev2", 00:33:07.667 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:07.667 "is_configured": false, 00:33:07.667 "data_offset": 0, 00:33:07.667 "data_size": 0 00:33:07.667 } 00:33:07.667 ] 00:33:07.667 }' 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:07.667 20:32:02 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:07.926 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2 00:33:07.926 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:07.926 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:08.186 [2024-10-01 20:32:03.221653] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:33:08.186 BaseBdev2 00:33:08.186 [2024-10-01 20:32:03.222103] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:33:08.186 [2024-10-01 20:32:03.222126] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:33:08.186 [2024-10-01 20:32:03.222334] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:33:08.186 [2024-10-01 20:32:03.222446] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:33:08.186 [2024-10-01 20:32:03.222472] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000007e80 00:33:08.186 [2024-10-01 20:32:03.222586] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@901 -- # local i 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:08.186 [ 00:33:08.186 { 00:33:08.186 "name": "BaseBdev2", 00:33:08.186 "aliases": [ 00:33:08.186 "128b4eb3-2186-4dda-b381-d0e5358e6874" 00:33:08.186 ], 00:33:08.186 "product_name": "Malloc disk", 00:33:08.186 "block_size": 4128, 00:33:08.186 "num_blocks": 8192, 00:33:08.186 "uuid": "128b4eb3-2186-4dda-b381-d0e5358e6874", 00:33:08.186 "md_size": 32, 00:33:08.186 "md_interleave": true, 00:33:08.186 "dif_type": 0, 00:33:08.186 "assigned_rate_limits": { 00:33:08.186 "rw_ios_per_sec": 0, 00:33:08.186 "rw_mbytes_per_sec": 0, 00:33:08.186 "r_mbytes_per_sec": 0, 00:33:08.186 "w_mbytes_per_sec": 0 00:33:08.186 }, 00:33:08.186 "claimed": true, 00:33:08.186 "claim_type": "exclusive_write", 00:33:08.186 "zoned": false, 00:33:08.186 "supported_io_types": { 00:33:08.186 "read": true, 00:33:08.186 "write": true, 00:33:08.186 "unmap": true, 00:33:08.186 "flush": true, 00:33:08.186 "reset": true, 00:33:08.186 "nvme_admin": false, 00:33:08.186 "nvme_io": false, 00:33:08.186 "nvme_io_md": false, 00:33:08.186 "write_zeroes": true, 00:33:08.186 "zcopy": true, 00:33:08.186 "get_zone_info": false, 00:33:08.186 "zone_management": false, 00:33:08.186 "zone_append": false, 00:33:08.186 "compare": false, 00:33:08.186 "compare_and_write": false, 00:33:08.186 "abort": true, 00:33:08.186 "seek_hole": false, 00:33:08.186 "seek_data": false, 00:33:08.186 "copy": true, 00:33:08.186 "nvme_iov_md": false 00:33:08.186 }, 00:33:08.186 "memory_domains": [ 00:33:08.186 { 00:33:08.186 "dma_device_id": "system", 00:33:08.186 "dma_device_type": 1 00:33:08.186 }, 00:33:08.186 { 00:33:08.186 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:08.186 "dma_device_type": 2 00:33:08.186 } 00:33:08.186 ], 00:33:08.186 "driver_specific": {} 00:33:08.186 } 00:33:08.186 ] 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@907 -- # return 0 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:08.186 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:08.186 "name": "Existed_Raid", 00:33:08.186 "uuid": "36bcab26-fbb4-4868-b384-5c99c5b1971a", 00:33:08.186 "strip_size_kb": 0, 00:33:08.186 "state": "online", 00:33:08.186 "raid_level": "raid1", 00:33:08.186 "superblock": true, 00:33:08.186 "num_base_bdevs": 2, 00:33:08.186 "num_base_bdevs_discovered": 2, 00:33:08.186 "num_base_bdevs_operational": 2, 00:33:08.186 "base_bdevs_list": [ 00:33:08.186 { 00:33:08.186 "name": "BaseBdev1", 00:33:08.187 "uuid": "b3f6c06c-3d04-42fb-bfc9-009915584700", 00:33:08.187 "is_configured": true, 00:33:08.187 "data_offset": 256, 00:33:08.187 "data_size": 7936 00:33:08.187 }, 00:33:08.187 { 00:33:08.187 "name": "BaseBdev2", 00:33:08.187 "uuid": "128b4eb3-2186-4dda-b381-d0e5358e6874", 00:33:08.187 "is_configured": true, 00:33:08.187 "data_offset": 256, 00:33:08.187 "data_size": 7936 00:33:08.187 } 00:33:08.187 ] 00:33:08.187 }' 00:33:08.187 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:08.187 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:08.755 [2024-10-01 20:32:03.798372] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:33:08.755 "name": "Existed_Raid", 00:33:08.755 "aliases": [ 00:33:08.755 "36bcab26-fbb4-4868-b384-5c99c5b1971a" 00:33:08.755 ], 00:33:08.755 "product_name": "Raid Volume", 00:33:08.755 "block_size": 4128, 00:33:08.755 "num_blocks": 7936, 00:33:08.755 "uuid": "36bcab26-fbb4-4868-b384-5c99c5b1971a", 00:33:08.755 "md_size": 32, 00:33:08.755 "md_interleave": true, 00:33:08.755 "dif_type": 0, 00:33:08.755 "assigned_rate_limits": { 00:33:08.755 "rw_ios_per_sec": 0, 00:33:08.755 "rw_mbytes_per_sec": 0, 00:33:08.755 "r_mbytes_per_sec": 0, 00:33:08.755 "w_mbytes_per_sec": 0 00:33:08.755 }, 00:33:08.755 "claimed": false, 00:33:08.755 "zoned": false, 00:33:08.755 "supported_io_types": { 00:33:08.755 "read": true, 00:33:08.755 "write": true, 00:33:08.755 "unmap": false, 00:33:08.755 "flush": false, 00:33:08.755 "reset": true, 00:33:08.755 "nvme_admin": false, 00:33:08.755 "nvme_io": false, 00:33:08.755 "nvme_io_md": false, 00:33:08.755 "write_zeroes": true, 00:33:08.755 "zcopy": false, 00:33:08.755 "get_zone_info": false, 00:33:08.755 "zone_management": false, 00:33:08.755 "zone_append": false, 00:33:08.755 "compare": false, 00:33:08.755 "compare_and_write": false, 00:33:08.755 "abort": false, 00:33:08.755 "seek_hole": false, 00:33:08.755 "seek_data": false, 00:33:08.755 "copy": false, 00:33:08.755 "nvme_iov_md": false 00:33:08.755 }, 00:33:08.755 "memory_domains": [ 00:33:08.755 { 00:33:08.755 "dma_device_id": "system", 00:33:08.755 "dma_device_type": 1 00:33:08.755 }, 00:33:08.755 { 00:33:08.755 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:08.755 "dma_device_type": 2 00:33:08.755 }, 00:33:08.755 { 00:33:08.755 "dma_device_id": "system", 00:33:08.755 "dma_device_type": 1 00:33:08.755 }, 00:33:08.755 { 00:33:08.755 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:08.755 "dma_device_type": 2 00:33:08.755 } 00:33:08.755 ], 00:33:08.755 "driver_specific": { 00:33:08.755 "raid": { 00:33:08.755 "uuid": "36bcab26-fbb4-4868-b384-5c99c5b1971a", 00:33:08.755 "strip_size_kb": 0, 00:33:08.755 "state": "online", 00:33:08.755 "raid_level": "raid1", 00:33:08.755 "superblock": true, 00:33:08.755 "num_base_bdevs": 2, 00:33:08.755 "num_base_bdevs_discovered": 2, 00:33:08.755 "num_base_bdevs_operational": 2, 00:33:08.755 "base_bdevs_list": [ 00:33:08.755 { 00:33:08.755 "name": "BaseBdev1", 00:33:08.755 "uuid": "b3f6c06c-3d04-42fb-bfc9-009915584700", 00:33:08.755 "is_configured": true, 00:33:08.755 "data_offset": 256, 00:33:08.755 "data_size": 7936 00:33:08.755 }, 00:33:08.755 { 00:33:08.755 "name": "BaseBdev2", 00:33:08.755 "uuid": "128b4eb3-2186-4dda-b381-d0e5358e6874", 00:33:08.755 "is_configured": true, 00:33:08.755 "data_offset": 256, 00:33:08.755 "data_size": 7936 00:33:08.755 } 00:33:08.755 ] 00:33:08.755 } 00:33:08.755 } 00:33:08.755 }' 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:33:08.755 BaseBdev2' 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:33:08.755 20:32:03 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:08.755 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:33:08.755 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:08.755 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:08.755 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:09.015 [2024-10-01 20:32:04.062071] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@260 -- # local expected_state 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:09.015 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:09.016 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:33:09.016 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:09.016 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:09.016 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:09.016 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:09.016 "name": "Existed_Raid", 00:33:09.016 "uuid": "36bcab26-fbb4-4868-b384-5c99c5b1971a", 00:33:09.016 "strip_size_kb": 0, 00:33:09.016 "state": "online", 00:33:09.016 "raid_level": "raid1", 00:33:09.016 "superblock": true, 00:33:09.016 "num_base_bdevs": 2, 00:33:09.016 "num_base_bdevs_discovered": 1, 00:33:09.016 "num_base_bdevs_operational": 1, 00:33:09.016 "base_bdevs_list": [ 00:33:09.016 { 00:33:09.016 "name": null, 00:33:09.016 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:09.016 "is_configured": false, 00:33:09.016 "data_offset": 0, 00:33:09.016 "data_size": 7936 00:33:09.016 }, 00:33:09.016 { 00:33:09.016 "name": "BaseBdev2", 00:33:09.016 "uuid": "128b4eb3-2186-4dda-b381-d0e5358e6874", 00:33:09.016 "is_configured": true, 00:33:09.016 "data_offset": 256, 00:33:09.016 "data_size": 7936 00:33:09.016 } 00:33:09.016 ] 00:33:09.016 }' 00:33:09.016 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:09.016 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:09.632 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:33:09.632 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:33:09.632 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:09.632 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:33:09.632 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:09.632 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:09.632 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:09.632 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:33:09.632 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:33:09.632 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:33:09.632 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:09.632 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:09.633 [2024-10-01 20:32:04.717250] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:33:09.633 [2024-10-01 20:32:04.717415] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:33:09.633 [2024-10-01 20:32:04.807341] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:09.633 [2024-10-01 20:32:04.807415] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:09.633 [2024-10-01 20:32:04.807440] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name Existed_Raid, state offline 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@326 -- # killprocess 89976 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@950 -- # '[' -z 89976 ']' 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # kill -0 89976 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@955 -- # uname 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:33:09.633 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 89976 00:33:09.892 killing process with pid 89976 00:33:09.892 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:33:09.892 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:33:09.892 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@968 -- # echo 'killing process with pid 89976' 00:33:09.892 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@969 -- # kill 89976 00:33:09.892 [2024-10-01 20:32:04.898596] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:33:09.892 20:32:04 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@974 -- # wait 89976 00:33:09.892 [2024-10-01 20:32:04.915841] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:33:11.798 20:32:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@328 -- # return 0 00:33:11.798 00:33:11.798 real 0m6.881s 00:33:11.798 user 0m9.561s 00:33:11.798 sys 0m1.064s 00:33:11.798 20:32:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1126 -- # xtrace_disable 00:33:11.798 20:32:06 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:11.798 ************************************ 00:33:11.798 END TEST raid_state_function_test_sb_md_interleaved 00:33:11.798 ************************************ 00:33:11.798 20:32:07 bdev_raid -- bdev/bdev_raid.sh@1012 -- # run_test raid_superblock_test_md_interleaved raid_superblock_test raid1 2 00:33:11.798 20:32:07 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:33:11.798 20:32:07 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:33:11.798 20:32:07 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:33:12.057 ************************************ 00:33:12.057 START TEST raid_superblock_test_md_interleaved 00:33:12.057 ************************************ 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 2 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@399 -- # local strip_size 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@412 -- # raid_pid=90243 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@413 -- # waitforlisten 90243 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@831 -- # '[' -z 90243 ']' 00:33:12.057 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@836 -- # local max_retries=100 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@840 -- # xtrace_disable 00:33:12.057 20:32:07 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:12.057 [2024-10-01 20:32:07.169163] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:33:12.057 [2024-10-01 20:32:07.169353] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid90243 ] 00:33:12.316 [2024-10-01 20:32:07.346611] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:12.575 [2024-10-01 20:32:07.647834] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:33:12.835 [2024-10-01 20:32:07.872361] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:33:12.835 [2024-10-01 20:32:07.872424] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:33:13.095 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:33:13.095 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@864 -- # return 0 00:33:13.095 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:33:13.095 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:33:13.095 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:33:13.095 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:33:13.095 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:33:13.095 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:33:13.095 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:33:13.095 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:33:13.095 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc1 00:33:13.095 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:13.095 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:13.355 malloc1 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:13.355 [2024-10-01 20:32:08.386923] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:33:13.355 [2024-10-01 20:32:08.386994] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:13.355 [2024-10-01 20:32:08.387036] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:33:13.355 [2024-10-01 20:32:08.387053] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:13.355 [2024-10-01 20:32:08.389910] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:13.355 [2024-10-01 20:32:08.390098] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:33:13.355 pt1 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc2 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:13.355 malloc2 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:13.355 [2024-10-01 20:32:08.447867] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:33:13.355 [2024-10-01 20:32:08.447966] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:13.355 [2024-10-01 20:32:08.448005] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:33:13.355 [2024-10-01 20:32:08.448023] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:13.355 [2024-10-01 20:32:08.450875] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:13.355 [2024-10-01 20:32:08.450936] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:33:13.355 pt2 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:13.355 [2024-10-01 20:32:08.459968] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:33:13.355 [2024-10-01 20:32:08.463036] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:33:13.355 [2024-10-01 20:32:08.463347] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:33:13.355 [2024-10-01 20:32:08.463367] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:33:13.355 [2024-10-01 20:32:08.463455] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005d40 00:33:13.355 [2024-10-01 20:32:08.463543] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:33:13.355 [2024-10-01 20:32:08.463566] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:33:13.355 [2024-10-01 20:32:08.463665] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:13.355 "name": "raid_bdev1", 00:33:13.355 "uuid": "3ab3d259-e703-43bb-8033-9b6ef8369061", 00:33:13.355 "strip_size_kb": 0, 00:33:13.355 "state": "online", 00:33:13.355 "raid_level": "raid1", 00:33:13.355 "superblock": true, 00:33:13.355 "num_base_bdevs": 2, 00:33:13.355 "num_base_bdevs_discovered": 2, 00:33:13.355 "num_base_bdevs_operational": 2, 00:33:13.355 "base_bdevs_list": [ 00:33:13.355 { 00:33:13.355 "name": "pt1", 00:33:13.355 "uuid": "00000000-0000-0000-0000-000000000001", 00:33:13.355 "is_configured": true, 00:33:13.355 "data_offset": 256, 00:33:13.355 "data_size": 7936 00:33:13.355 }, 00:33:13.355 { 00:33:13.355 "name": "pt2", 00:33:13.355 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:13.355 "is_configured": true, 00:33:13.355 "data_offset": 256, 00:33:13.355 "data_size": 7936 00:33:13.355 } 00:33:13.355 ] 00:33:13.355 }' 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:13.355 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:13.935 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:33:13.935 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:33:13.935 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:33:13.935 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:33:13.935 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:33:13.935 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:33:13.935 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:33:13.935 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:13.935 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:13.935 20:32:08 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:33:13.935 [2024-10-01 20:32:08.992692] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:13.935 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:13.935 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:33:13.935 "name": "raid_bdev1", 00:33:13.935 "aliases": [ 00:33:13.936 "3ab3d259-e703-43bb-8033-9b6ef8369061" 00:33:13.936 ], 00:33:13.936 "product_name": "Raid Volume", 00:33:13.936 "block_size": 4128, 00:33:13.936 "num_blocks": 7936, 00:33:13.936 "uuid": "3ab3d259-e703-43bb-8033-9b6ef8369061", 00:33:13.936 "md_size": 32, 00:33:13.936 "md_interleave": true, 00:33:13.936 "dif_type": 0, 00:33:13.936 "assigned_rate_limits": { 00:33:13.936 "rw_ios_per_sec": 0, 00:33:13.936 "rw_mbytes_per_sec": 0, 00:33:13.936 "r_mbytes_per_sec": 0, 00:33:13.936 "w_mbytes_per_sec": 0 00:33:13.936 }, 00:33:13.936 "claimed": false, 00:33:13.936 "zoned": false, 00:33:13.936 "supported_io_types": { 00:33:13.936 "read": true, 00:33:13.936 "write": true, 00:33:13.936 "unmap": false, 00:33:13.936 "flush": false, 00:33:13.936 "reset": true, 00:33:13.936 "nvme_admin": false, 00:33:13.936 "nvme_io": false, 00:33:13.936 "nvme_io_md": false, 00:33:13.936 "write_zeroes": true, 00:33:13.936 "zcopy": false, 00:33:13.936 "get_zone_info": false, 00:33:13.936 "zone_management": false, 00:33:13.936 "zone_append": false, 00:33:13.936 "compare": false, 00:33:13.936 "compare_and_write": false, 00:33:13.936 "abort": false, 00:33:13.936 "seek_hole": false, 00:33:13.936 "seek_data": false, 00:33:13.936 "copy": false, 00:33:13.936 "nvme_iov_md": false 00:33:13.936 }, 00:33:13.936 "memory_domains": [ 00:33:13.936 { 00:33:13.936 "dma_device_id": "system", 00:33:13.936 "dma_device_type": 1 00:33:13.936 }, 00:33:13.936 { 00:33:13.936 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:13.936 "dma_device_type": 2 00:33:13.936 }, 00:33:13.936 { 00:33:13.936 "dma_device_id": "system", 00:33:13.936 "dma_device_type": 1 00:33:13.936 }, 00:33:13.936 { 00:33:13.936 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:13.936 "dma_device_type": 2 00:33:13.936 } 00:33:13.936 ], 00:33:13.936 "driver_specific": { 00:33:13.936 "raid": { 00:33:13.936 "uuid": "3ab3d259-e703-43bb-8033-9b6ef8369061", 00:33:13.936 "strip_size_kb": 0, 00:33:13.936 "state": "online", 00:33:13.936 "raid_level": "raid1", 00:33:13.936 "superblock": true, 00:33:13.936 "num_base_bdevs": 2, 00:33:13.936 "num_base_bdevs_discovered": 2, 00:33:13.936 "num_base_bdevs_operational": 2, 00:33:13.936 "base_bdevs_list": [ 00:33:13.936 { 00:33:13.936 "name": "pt1", 00:33:13.936 "uuid": "00000000-0000-0000-0000-000000000001", 00:33:13.936 "is_configured": true, 00:33:13.936 "data_offset": 256, 00:33:13.936 "data_size": 7936 00:33:13.936 }, 00:33:13.936 { 00:33:13.936 "name": "pt2", 00:33:13.936 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:13.936 "is_configured": true, 00:33:13.936 "data_offset": 256, 00:33:13.936 "data_size": 7936 00:33:13.936 } 00:33:13.936 ] 00:33:13.936 } 00:33:13.936 } 00:33:13.936 }' 00:33:13.936 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:33:13.936 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:33:13.936 pt2' 00:33:13.936 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:13.936 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:33:13.936 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:13.936 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:33:13.936 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:13.936 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:13.936 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:13.936 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:14.202 [2024-10-01 20:32:09.268837] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=3ab3d259-e703-43bb-8033-9b6ef8369061 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@436 -- # '[' -z 3ab3d259-e703-43bb-8033-9b6ef8369061 ']' 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:14.202 [2024-10-01 20:32:09.320416] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:33:14.202 [2024-10-01 20:32:09.320449] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:33:14.202 [2024-10-01 20:32:09.320550] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:14.202 [2024-10-01 20:32:09.320640] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:14.202 [2024-10-01 20:32:09.320667] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:33:14.202 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@650 -- # local es=0 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:14.463 [2024-10-01 20:32:09.464500] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:33:14.463 [2024-10-01 20:32:09.467388] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:33:14.463 [2024-10-01 20:32:09.467514] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:33:14.463 [2024-10-01 20:32:09.467620] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:33:14.463 [2024-10-01 20:32:09.467654] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:33:14.463 [2024-10-01 20:32:09.467676] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state configuring 00:33:14.463 request: 00:33:14.463 { 00:33:14.463 "name": "raid_bdev1", 00:33:14.463 "raid_level": "raid1", 00:33:14.463 "base_bdevs": [ 00:33:14.463 "malloc1", 00:33:14.463 "malloc2" 00:33:14.463 ], 00:33:14.463 "superblock": false, 00:33:14.463 "method": "bdev_raid_create", 00:33:14.463 "req_id": 1 00:33:14.463 } 00:33:14.463 Got JSON-RPC error response 00:33:14.463 response: 00:33:14.463 { 00:33:14.463 "code": -17, 00:33:14.463 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:33:14.463 } 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@653 -- # es=1 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:14.463 [2024-10-01 20:32:09.536602] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:33:14.463 [2024-10-01 20:32:09.536700] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:14.463 [2024-10-01 20:32:09.536772] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:33:14.463 [2024-10-01 20:32:09.536825] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:14.463 [2024-10-01 20:32:09.539905] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:14.463 [2024-10-01 20:32:09.539951] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:33:14.463 [2024-10-01 20:32:09.540022] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:33:14.463 [2024-10-01 20:32:09.540121] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:33:14.463 pt1 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:14.463 "name": "raid_bdev1", 00:33:14.463 "uuid": "3ab3d259-e703-43bb-8033-9b6ef8369061", 00:33:14.463 "strip_size_kb": 0, 00:33:14.463 "state": "configuring", 00:33:14.463 "raid_level": "raid1", 00:33:14.463 "superblock": true, 00:33:14.463 "num_base_bdevs": 2, 00:33:14.463 "num_base_bdevs_discovered": 1, 00:33:14.463 "num_base_bdevs_operational": 2, 00:33:14.463 "base_bdevs_list": [ 00:33:14.463 { 00:33:14.463 "name": "pt1", 00:33:14.463 "uuid": "00000000-0000-0000-0000-000000000001", 00:33:14.463 "is_configured": true, 00:33:14.463 "data_offset": 256, 00:33:14.463 "data_size": 7936 00:33:14.463 }, 00:33:14.463 { 00:33:14.463 "name": null, 00:33:14.463 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:14.463 "is_configured": false, 00:33:14.463 "data_offset": 256, 00:33:14.463 "data_size": 7936 00:33:14.463 } 00:33:14.463 ] 00:33:14.463 }' 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:14.463 20:32:09 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:15.032 [2024-10-01 20:32:10.092907] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:33:15.032 [2024-10-01 20:32:10.093002] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:15.032 [2024-10-01 20:32:10.093040] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:33:15.032 [2024-10-01 20:32:10.093060] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:15.032 [2024-10-01 20:32:10.093339] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:15.032 [2024-10-01 20:32:10.093369] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:33:15.032 [2024-10-01 20:32:10.093502] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:33:15.032 [2024-10-01 20:32:10.093544] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:33:15.032 [2024-10-01 20:32:10.093658] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007e80 00:33:15.032 [2024-10-01 20:32:10.093688] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:33:15.032 [2024-10-01 20:32:10.093803] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:33:15.032 [2024-10-01 20:32:10.093895] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007e80 00:33:15.032 [2024-10-01 20:32:10.093910] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007e80 00:33:15.032 [2024-10-01 20:32:10.094014] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:15.032 pt2 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:15.032 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:15.033 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:15.033 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:15.033 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:15.033 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:15.033 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:15.033 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:15.033 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:15.033 "name": "raid_bdev1", 00:33:15.033 "uuid": "3ab3d259-e703-43bb-8033-9b6ef8369061", 00:33:15.033 "strip_size_kb": 0, 00:33:15.033 "state": "online", 00:33:15.033 "raid_level": "raid1", 00:33:15.033 "superblock": true, 00:33:15.033 "num_base_bdevs": 2, 00:33:15.033 "num_base_bdevs_discovered": 2, 00:33:15.033 "num_base_bdevs_operational": 2, 00:33:15.033 "base_bdevs_list": [ 00:33:15.033 { 00:33:15.033 "name": "pt1", 00:33:15.033 "uuid": "00000000-0000-0000-0000-000000000001", 00:33:15.033 "is_configured": true, 00:33:15.033 "data_offset": 256, 00:33:15.033 "data_size": 7936 00:33:15.033 }, 00:33:15.033 { 00:33:15.033 "name": "pt2", 00:33:15.033 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:15.033 "is_configured": true, 00:33:15.033 "data_offset": 256, 00:33:15.033 "data_size": 7936 00:33:15.033 } 00:33:15.033 ] 00:33:15.033 }' 00:33:15.033 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:15.033 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:15.601 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:33:15.601 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:33:15.601 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:33:15.601 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:33:15.601 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:33:15.601 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:33:15.601 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:33:15.601 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:15.601 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:15.601 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:33:15.601 [2024-10-01 20:32:10.637519] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:15.601 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:15.601 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:33:15.601 "name": "raid_bdev1", 00:33:15.601 "aliases": [ 00:33:15.601 "3ab3d259-e703-43bb-8033-9b6ef8369061" 00:33:15.601 ], 00:33:15.601 "product_name": "Raid Volume", 00:33:15.601 "block_size": 4128, 00:33:15.601 "num_blocks": 7936, 00:33:15.601 "uuid": "3ab3d259-e703-43bb-8033-9b6ef8369061", 00:33:15.601 "md_size": 32, 00:33:15.601 "md_interleave": true, 00:33:15.601 "dif_type": 0, 00:33:15.601 "assigned_rate_limits": { 00:33:15.601 "rw_ios_per_sec": 0, 00:33:15.601 "rw_mbytes_per_sec": 0, 00:33:15.601 "r_mbytes_per_sec": 0, 00:33:15.601 "w_mbytes_per_sec": 0 00:33:15.601 }, 00:33:15.601 "claimed": false, 00:33:15.602 "zoned": false, 00:33:15.602 "supported_io_types": { 00:33:15.602 "read": true, 00:33:15.602 "write": true, 00:33:15.602 "unmap": false, 00:33:15.602 "flush": false, 00:33:15.602 "reset": true, 00:33:15.602 "nvme_admin": false, 00:33:15.602 "nvme_io": false, 00:33:15.602 "nvme_io_md": false, 00:33:15.602 "write_zeroes": true, 00:33:15.602 "zcopy": false, 00:33:15.602 "get_zone_info": false, 00:33:15.602 "zone_management": false, 00:33:15.602 "zone_append": false, 00:33:15.602 "compare": false, 00:33:15.602 "compare_and_write": false, 00:33:15.602 "abort": false, 00:33:15.602 "seek_hole": false, 00:33:15.602 "seek_data": false, 00:33:15.602 "copy": false, 00:33:15.602 "nvme_iov_md": false 00:33:15.602 }, 00:33:15.602 "memory_domains": [ 00:33:15.602 { 00:33:15.602 "dma_device_id": "system", 00:33:15.602 "dma_device_type": 1 00:33:15.602 }, 00:33:15.602 { 00:33:15.602 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:15.602 "dma_device_type": 2 00:33:15.602 }, 00:33:15.602 { 00:33:15.602 "dma_device_id": "system", 00:33:15.602 "dma_device_type": 1 00:33:15.602 }, 00:33:15.602 { 00:33:15.602 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:33:15.602 "dma_device_type": 2 00:33:15.602 } 00:33:15.602 ], 00:33:15.602 "driver_specific": { 00:33:15.602 "raid": { 00:33:15.602 "uuid": "3ab3d259-e703-43bb-8033-9b6ef8369061", 00:33:15.602 "strip_size_kb": 0, 00:33:15.602 "state": "online", 00:33:15.602 "raid_level": "raid1", 00:33:15.602 "superblock": true, 00:33:15.602 "num_base_bdevs": 2, 00:33:15.602 "num_base_bdevs_discovered": 2, 00:33:15.602 "num_base_bdevs_operational": 2, 00:33:15.602 "base_bdevs_list": [ 00:33:15.602 { 00:33:15.602 "name": "pt1", 00:33:15.602 "uuid": "00000000-0000-0000-0000-000000000001", 00:33:15.602 "is_configured": true, 00:33:15.602 "data_offset": 256, 00:33:15.602 "data_size": 7936 00:33:15.602 }, 00:33:15.602 { 00:33:15.602 "name": "pt2", 00:33:15.602 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:15.602 "is_configured": true, 00:33:15.602 "data_offset": 256, 00:33:15.602 "data_size": 7936 00:33:15.602 } 00:33:15.602 ] 00:33:15.602 } 00:33:15.602 } 00:33:15.602 }' 00:33:15.602 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:33:15.602 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:33:15.602 pt2' 00:33:15.602 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:15.602 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:33:15.602 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:15.602 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:33:15.602 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:15.602 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:15.602 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:15.602 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:15.602 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:33:15.602 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:33:15.602 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:33:15.602 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:33:15.862 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:33:15.862 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:15.862 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:15.862 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:15.862 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:33:15.862 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:33:15.862 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:33:15.862 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:15.862 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:33:15.862 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:15.862 [2024-10-01 20:32:10.917554] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # '[' 3ab3d259-e703-43bb-8033-9b6ef8369061 '!=' 3ab3d259-e703-43bb-8033-9b6ef8369061 ']' 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:15.863 [2024-10-01 20:32:10.965303] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:15.863 20:32:10 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:15.863 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:15.863 "name": "raid_bdev1", 00:33:15.863 "uuid": "3ab3d259-e703-43bb-8033-9b6ef8369061", 00:33:15.863 "strip_size_kb": 0, 00:33:15.863 "state": "online", 00:33:15.863 "raid_level": "raid1", 00:33:15.863 "superblock": true, 00:33:15.863 "num_base_bdevs": 2, 00:33:15.863 "num_base_bdevs_discovered": 1, 00:33:15.863 "num_base_bdevs_operational": 1, 00:33:15.863 "base_bdevs_list": [ 00:33:15.863 { 00:33:15.863 "name": null, 00:33:15.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:15.863 "is_configured": false, 00:33:15.863 "data_offset": 0, 00:33:15.863 "data_size": 7936 00:33:15.863 }, 00:33:15.863 { 00:33:15.863 "name": "pt2", 00:33:15.863 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:15.863 "is_configured": true, 00:33:15.863 "data_offset": 256, 00:33:15.863 "data_size": 7936 00:33:15.863 } 00:33:15.863 ] 00:33:15.863 }' 00:33:15.863 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:15.863 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:16.433 [2024-10-01 20:32:11.509501] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:33:16.433 [2024-10-01 20:32:11.509539] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:33:16.433 [2024-10-01 20:32:11.509644] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:16.433 [2024-10-01 20:32:11.509709] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:16.433 [2024-10-01 20:32:11.509754] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007e80 name raid_bdev1, state offline 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@519 -- # i=1 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:16.433 [2024-10-01 20:32:11.585531] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:33:16.433 [2024-10-01 20:32:11.585612] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:16.433 [2024-10-01 20:32:11.585641] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:33:16.433 [2024-10-01 20:32:11.585661] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:16.433 [2024-10-01 20:32:11.588846] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:16.433 [2024-10-01 20:32:11.589011] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:33:16.433 [2024-10-01 20:32:11.589195] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:33:16.433 [2024-10-01 20:32:11.589370] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:33:16.433 [2024-10-01 20:32:11.589503] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008200 00:33:16.433 [2024-10-01 20:32:11.589528] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:33:16.433 [2024-10-01 20:32:11.589655] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:33:16.433 [2024-10-01 20:32:11.589895] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008200 00:33:16.433 [2024-10-01 20:32:11.590019] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008200 00:33:16.433 [2024-10-01 20:32:11.590294] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:16.433 pt2 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:16.433 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:16.433 "name": "raid_bdev1", 00:33:16.433 "uuid": "3ab3d259-e703-43bb-8033-9b6ef8369061", 00:33:16.433 "strip_size_kb": 0, 00:33:16.433 "state": "online", 00:33:16.433 "raid_level": "raid1", 00:33:16.433 "superblock": true, 00:33:16.433 "num_base_bdevs": 2, 00:33:16.433 "num_base_bdevs_discovered": 1, 00:33:16.433 "num_base_bdevs_operational": 1, 00:33:16.433 "base_bdevs_list": [ 00:33:16.434 { 00:33:16.434 "name": null, 00:33:16.434 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:16.434 "is_configured": false, 00:33:16.434 "data_offset": 256, 00:33:16.434 "data_size": 7936 00:33:16.434 }, 00:33:16.434 { 00:33:16.434 "name": "pt2", 00:33:16.434 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:16.434 "is_configured": true, 00:33:16.434 "data_offset": 256, 00:33:16.434 "data_size": 7936 00:33:16.434 } 00:33:16.434 ] 00:33:16.434 }' 00:33:16.434 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:16.434 20:32:11 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:17.029 [2024-10-01 20:32:12.109794] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:33:17.029 [2024-10-01 20:32:12.109834] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:33:17.029 [2024-10-01 20:32:12.109962] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:17.029 [2024-10-01 20:32:12.110087] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:17.029 [2024-10-01 20:32:12.110103] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008200 name raid_bdev1, state offline 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:17.029 [2024-10-01 20:32:12.169875] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:33:17.029 [2024-10-01 20:32:12.170129] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:17.029 [2024-10-01 20:32:12.170186] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:33:17.029 [2024-10-01 20:32:12.170204] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:17.029 [2024-10-01 20:32:12.173362] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:17.029 [2024-10-01 20:32:12.173532] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:33:17.029 [2024-10-01 20:32:12.173626] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:33:17.029 [2024-10-01 20:32:12.173699] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:33:17.029 [2024-10-01 20:32:12.173850] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:33:17.029 [2024-10-01 20:32:12.173869] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:33:17.029 [2024-10-01 20:32:12.173899] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008580 name raid_bdev1, state configuring 00:33:17.029 [2024-10-01 20:32:12.173976] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:33:17.029 pt1 00:33:17.029 [2024-10-01 20:32:12.174136] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000008900 00:33:17.029 [2024-10-01 20:32:12.174158] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:33:17.029 [2024-10-01 20:32:12.174264] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:33:17.029 [2024-10-01 20:32:12.174376] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000008900 00:33:17.029 [2024-10-01 20:32:12.174396] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000008900 00:33:17.029 [2024-10-01 20:32:12.174495] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:17.029 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:17.029 "name": "raid_bdev1", 00:33:17.029 "uuid": "3ab3d259-e703-43bb-8033-9b6ef8369061", 00:33:17.030 "strip_size_kb": 0, 00:33:17.030 "state": "online", 00:33:17.030 "raid_level": "raid1", 00:33:17.030 "superblock": true, 00:33:17.030 "num_base_bdevs": 2, 00:33:17.030 "num_base_bdevs_discovered": 1, 00:33:17.030 "num_base_bdevs_operational": 1, 00:33:17.030 "base_bdevs_list": [ 00:33:17.030 { 00:33:17.030 "name": null, 00:33:17.030 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:17.030 "is_configured": false, 00:33:17.030 "data_offset": 256, 00:33:17.030 "data_size": 7936 00:33:17.030 }, 00:33:17.030 { 00:33:17.030 "name": "pt2", 00:33:17.030 "uuid": "00000000-0000-0000-0000-000000000002", 00:33:17.030 "is_configured": true, 00:33:17.030 "data_offset": 256, 00:33:17.030 "data_size": 7936 00:33:17.030 } 00:33:17.030 ] 00:33:17.030 }' 00:33:17.030 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:17.030 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:17.598 [2024-10-01 20:32:12.762642] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # '[' 3ab3d259-e703-43bb-8033-9b6ef8369061 '!=' 3ab3d259-e703-43bb-8033-9b6ef8369061 ']' 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@563 -- # killprocess 90243 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@950 -- # '[' -z 90243 ']' 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@954 -- # kill -0 90243 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@955 -- # uname 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 90243 00:33:17.598 killing process with pid 90243 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@968 -- # echo 'killing process with pid 90243' 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@969 -- # kill 90243 00:33:17.598 [2024-10-01 20:32:12.838926] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:33:17.598 20:32:12 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@974 -- # wait 90243 00:33:17.598 [2024-10-01 20:32:12.839050] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:17.598 [2024-10-01 20:32:12.839117] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:17.598 [2024-10-01 20:32:12.839157] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000008900 name raid_bdev1, state offline 00:33:17.857 [2024-10-01 20:32:13.044194] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:33:19.760 20:32:14 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@565 -- # return 0 00:33:19.760 00:33:19.760 real 0m7.939s 00:33:19.760 user 0m11.811s 00:33:19.760 sys 0m1.201s 00:33:19.760 20:32:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1126 -- # xtrace_disable 00:33:19.760 20:32:14 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:19.760 ************************************ 00:33:19.760 END TEST raid_superblock_test_md_interleaved 00:33:19.760 ************************************ 00:33:20.020 20:32:15 bdev_raid -- bdev/bdev_raid.sh@1013 -- # run_test raid_rebuild_test_sb_md_interleaved raid_rebuild_test raid1 2 true false false 00:33:20.020 20:32:15 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:33:20.020 20:32:15 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:33:20.020 20:32:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:33:20.020 ************************************ 00:33:20.020 START TEST raid_rebuild_test_sb_md_interleaved 00:33:20.020 ************************************ 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true false false 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # local verify=false 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # local strip_size 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@577 -- # local create_arg 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@579 -- # local data_offset 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@597 -- # raid_pid=90583 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@598 -- # waitforlisten 90583 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@831 -- # '[' -z 90583 ']' 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:20.020 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@836 -- # local max_retries=100 00:33:20.021 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:20.021 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:20.021 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # xtrace_disable 00:33:20.021 20:32:15 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:20.021 [2024-10-01 20:32:15.152436] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:33:20.021 [2024-10-01 20:32:15.152941] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid90583 ] 00:33:20.021 I/O size of 3145728 is greater than zero copy threshold (65536). 00:33:20.021 Zero copy mechanism will not be used. 00:33:20.279 [2024-10-01 20:32:15.334544] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:20.538 [2024-10-01 20:32:15.617662] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:33:20.798 [2024-10-01 20:32:15.848710] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:33:20.798 [2024-10-01 20:32:15.849104] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:33:21.058 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:33:21.058 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # return 0 00:33:21.058 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:33:21.058 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1_malloc 00:33:21.058 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:21.058 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:21.317 BaseBdev1_malloc 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:21.317 [2024-10-01 20:32:16.364543] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:33:21.317 [2024-10-01 20:32:16.364633] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:21.317 [2024-10-01 20:32:16.364670] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:33:21.317 [2024-10-01 20:32:16.364689] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:21.317 [2024-10-01 20:32:16.367744] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:21.317 [2024-10-01 20:32:16.367814] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:33:21.317 BaseBdev1 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2_malloc 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:21.317 BaseBdev2_malloc 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:21.317 [2024-10-01 20:32:16.425537] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:33:21.317 [2024-10-01 20:32:16.425656] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:21.317 [2024-10-01 20:32:16.425703] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:33:21.317 [2024-10-01 20:32:16.425737] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:21.317 [2024-10-01 20:32:16.428817] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:21.317 [2024-10-01 20:32:16.428917] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:33:21.317 BaseBdev2 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b spare_malloc 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:21.317 spare_malloc 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:21.317 spare_delay 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:21.317 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:21.318 [2024-10-01 20:32:16.497056] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:33:21.318 [2024-10-01 20:32:16.497279] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:21.318 [2024-10-01 20:32:16.497456] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:33:21.318 [2024-10-01 20:32:16.497577] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:21.318 [2024-10-01 20:32:16.500450] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:21.318 [2024-10-01 20:32:16.500619] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:33:21.318 spare 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:21.318 [2024-10-01 20:32:16.509449] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:33:21.318 [2024-10-01 20:32:16.512792] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:33:21.318 [2024-10-01 20:32:16.513244] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007780 00:33:21.318 [2024-10-01 20:32:16.513384] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:33:21.318 [2024-10-01 20:32:16.513531] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005e10 00:33:21.318 [2024-10-01 20:32:16.513638] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007780 00:33:21.318 [2024-10-01 20:32:16.513654] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007780 00:33:21.318 [2024-10-01 20:32:16.513810] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:21.318 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:21.577 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:21.577 "name": "raid_bdev1", 00:33:21.577 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:21.577 "strip_size_kb": 0, 00:33:21.577 "state": "online", 00:33:21.577 "raid_level": "raid1", 00:33:21.577 "superblock": true, 00:33:21.577 "num_base_bdevs": 2, 00:33:21.577 "num_base_bdevs_discovered": 2, 00:33:21.577 "num_base_bdevs_operational": 2, 00:33:21.577 "base_bdevs_list": [ 00:33:21.577 { 00:33:21.577 "name": "BaseBdev1", 00:33:21.577 "uuid": "e0de24f2-d15f-5066-94d2-7a7c9a954394", 00:33:21.577 "is_configured": true, 00:33:21.577 "data_offset": 256, 00:33:21.577 "data_size": 7936 00:33:21.577 }, 00:33:21.577 { 00:33:21.577 "name": "BaseBdev2", 00:33:21.577 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:21.577 "is_configured": true, 00:33:21.577 "data_offset": 256, 00:33:21.577 "data_size": 7936 00:33:21.577 } 00:33:21.577 ] 00:33:21.577 }' 00:33:21.577 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:21.577 20:32:16 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:21.836 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:33:21.836 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:21.836 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:21.836 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:33:21.836 [2024-10-01 20:32:17.018169] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:33:21.836 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:21.836 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:33:21.836 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:21.836 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:21.836 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:21.836 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:33:21.836 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@624 -- # '[' false = true ']' 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:22.095 [2024-10-01 20:32:17.125757] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:22.095 "name": "raid_bdev1", 00:33:22.095 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:22.095 "strip_size_kb": 0, 00:33:22.095 "state": "online", 00:33:22.095 "raid_level": "raid1", 00:33:22.095 "superblock": true, 00:33:22.095 "num_base_bdevs": 2, 00:33:22.095 "num_base_bdevs_discovered": 1, 00:33:22.095 "num_base_bdevs_operational": 1, 00:33:22.095 "base_bdevs_list": [ 00:33:22.095 { 00:33:22.095 "name": null, 00:33:22.095 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:22.095 "is_configured": false, 00:33:22.095 "data_offset": 0, 00:33:22.095 "data_size": 7936 00:33:22.095 }, 00:33:22.095 { 00:33:22.095 "name": "BaseBdev2", 00:33:22.095 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:22.095 "is_configured": true, 00:33:22.095 "data_offset": 256, 00:33:22.095 "data_size": 7936 00:33:22.095 } 00:33:22.095 ] 00:33:22.095 }' 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:22.095 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:22.663 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:33:22.663 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:22.663 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:22.663 [2024-10-01 20:32:17.642041] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:33:22.663 [2024-10-01 20:32:17.660233] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005ee0 00:33:22.663 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:22.663 20:32:17 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@647 -- # sleep 1 00:33:22.663 [2024-10-01 20:32:17.663146] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:23.608 "name": "raid_bdev1", 00:33:23.608 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:23.608 "strip_size_kb": 0, 00:33:23.608 "state": "online", 00:33:23.608 "raid_level": "raid1", 00:33:23.608 "superblock": true, 00:33:23.608 "num_base_bdevs": 2, 00:33:23.608 "num_base_bdevs_discovered": 2, 00:33:23.608 "num_base_bdevs_operational": 2, 00:33:23.608 "process": { 00:33:23.608 "type": "rebuild", 00:33:23.608 "target": "spare", 00:33:23.608 "progress": { 00:33:23.608 "blocks": 2560, 00:33:23.608 "percent": 32 00:33:23.608 } 00:33:23.608 }, 00:33:23.608 "base_bdevs_list": [ 00:33:23.608 { 00:33:23.608 "name": "spare", 00:33:23.608 "uuid": "dcc4ea57-1a06-5d7b-9fab-2c0c863ff1b2", 00:33:23.608 "is_configured": true, 00:33:23.608 "data_offset": 256, 00:33:23.608 "data_size": 7936 00:33:23.608 }, 00:33:23.608 { 00:33:23.608 "name": "BaseBdev2", 00:33:23.608 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:23.608 "is_configured": true, 00:33:23.608 "data_offset": 256, 00:33:23.608 "data_size": 7936 00:33:23.608 } 00:33:23.608 ] 00:33:23.608 }' 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:23.608 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:23.608 [2024-10-01 20:32:18.844755] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:33:23.878 [2024-10-01 20:32:18.873436] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:33:23.878 [2024-10-01 20:32:18.873527] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:23.878 [2024-10-01 20:32:18.873551] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:33:23.878 [2024-10-01 20:32:18.873567] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:23.878 "name": "raid_bdev1", 00:33:23.878 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:23.878 "strip_size_kb": 0, 00:33:23.878 "state": "online", 00:33:23.878 "raid_level": "raid1", 00:33:23.878 "superblock": true, 00:33:23.878 "num_base_bdevs": 2, 00:33:23.878 "num_base_bdevs_discovered": 1, 00:33:23.878 "num_base_bdevs_operational": 1, 00:33:23.878 "base_bdevs_list": [ 00:33:23.878 { 00:33:23.878 "name": null, 00:33:23.878 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:23.878 "is_configured": false, 00:33:23.878 "data_offset": 0, 00:33:23.878 "data_size": 7936 00:33:23.878 }, 00:33:23.878 { 00:33:23.878 "name": "BaseBdev2", 00:33:23.878 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:23.878 "is_configured": true, 00:33:23.878 "data_offset": 256, 00:33:23.878 "data_size": 7936 00:33:23.878 } 00:33:23.878 ] 00:33:23.878 }' 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:23.878 20:32:18 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:24.446 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:24.447 "name": "raid_bdev1", 00:33:24.447 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:24.447 "strip_size_kb": 0, 00:33:24.447 "state": "online", 00:33:24.447 "raid_level": "raid1", 00:33:24.447 "superblock": true, 00:33:24.447 "num_base_bdevs": 2, 00:33:24.447 "num_base_bdevs_discovered": 1, 00:33:24.447 "num_base_bdevs_operational": 1, 00:33:24.447 "base_bdevs_list": [ 00:33:24.447 { 00:33:24.447 "name": null, 00:33:24.447 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:24.447 "is_configured": false, 00:33:24.447 "data_offset": 0, 00:33:24.447 "data_size": 7936 00:33:24.447 }, 00:33:24.447 { 00:33:24.447 "name": "BaseBdev2", 00:33:24.447 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:24.447 "is_configured": true, 00:33:24.447 "data_offset": 256, 00:33:24.447 "data_size": 7936 00:33:24.447 } 00:33:24.447 ] 00:33:24.447 }' 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:24.447 [2024-10-01 20:32:19.616325] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:33:24.447 [2024-10-01 20:32:19.633349] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000005fb0 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:24.447 20:32:19 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@663 -- # sleep 1 00:33:24.447 [2024-10-01 20:32:19.636260] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:25.826 "name": "raid_bdev1", 00:33:25.826 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:25.826 "strip_size_kb": 0, 00:33:25.826 "state": "online", 00:33:25.826 "raid_level": "raid1", 00:33:25.826 "superblock": true, 00:33:25.826 "num_base_bdevs": 2, 00:33:25.826 "num_base_bdevs_discovered": 2, 00:33:25.826 "num_base_bdevs_operational": 2, 00:33:25.826 "process": { 00:33:25.826 "type": "rebuild", 00:33:25.826 "target": "spare", 00:33:25.826 "progress": { 00:33:25.826 "blocks": 2560, 00:33:25.826 "percent": 32 00:33:25.826 } 00:33:25.826 }, 00:33:25.826 "base_bdevs_list": [ 00:33:25.826 { 00:33:25.826 "name": "spare", 00:33:25.826 "uuid": "dcc4ea57-1a06-5d7b-9fab-2c0c863ff1b2", 00:33:25.826 "is_configured": true, 00:33:25.826 "data_offset": 256, 00:33:25.826 "data_size": 7936 00:33:25.826 }, 00:33:25.826 { 00:33:25.826 "name": "BaseBdev2", 00:33:25.826 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:25.826 "is_configured": true, 00:33:25.826 "data_offset": 256, 00:33:25.826 "data_size": 7936 00:33:25.826 } 00:33:25.826 ] 00:33:25.826 }' 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:33:25.826 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:33:25.826 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@706 -- # local timeout=879 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:25.827 "name": "raid_bdev1", 00:33:25.827 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:25.827 "strip_size_kb": 0, 00:33:25.827 "state": "online", 00:33:25.827 "raid_level": "raid1", 00:33:25.827 "superblock": true, 00:33:25.827 "num_base_bdevs": 2, 00:33:25.827 "num_base_bdevs_discovered": 2, 00:33:25.827 "num_base_bdevs_operational": 2, 00:33:25.827 "process": { 00:33:25.827 "type": "rebuild", 00:33:25.827 "target": "spare", 00:33:25.827 "progress": { 00:33:25.827 "blocks": 2816, 00:33:25.827 "percent": 35 00:33:25.827 } 00:33:25.827 }, 00:33:25.827 "base_bdevs_list": [ 00:33:25.827 { 00:33:25.827 "name": "spare", 00:33:25.827 "uuid": "dcc4ea57-1a06-5d7b-9fab-2c0c863ff1b2", 00:33:25.827 "is_configured": true, 00:33:25.827 "data_offset": 256, 00:33:25.827 "data_size": 7936 00:33:25.827 }, 00:33:25.827 { 00:33:25.827 "name": "BaseBdev2", 00:33:25.827 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:25.827 "is_configured": true, 00:33:25.827 "data_offset": 256, 00:33:25.827 "data_size": 7936 00:33:25.827 } 00:33:25.827 ] 00:33:25.827 }' 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:25.827 20:32:20 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:33:26.765 20:32:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:26.765 20:32:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:26.765 20:32:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:26.765 20:32:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:26.765 20:32:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:26.765 20:32:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:26.765 20:32:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:26.765 20:32:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:26.765 20:32:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:26.765 20:32:21 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:26.765 20:32:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:27.024 20:32:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:27.024 "name": "raid_bdev1", 00:33:27.024 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:27.024 "strip_size_kb": 0, 00:33:27.024 "state": "online", 00:33:27.024 "raid_level": "raid1", 00:33:27.024 "superblock": true, 00:33:27.024 "num_base_bdevs": 2, 00:33:27.024 "num_base_bdevs_discovered": 2, 00:33:27.024 "num_base_bdevs_operational": 2, 00:33:27.024 "process": { 00:33:27.024 "type": "rebuild", 00:33:27.024 "target": "spare", 00:33:27.024 "progress": { 00:33:27.024 "blocks": 5888, 00:33:27.024 "percent": 74 00:33:27.024 } 00:33:27.024 }, 00:33:27.024 "base_bdevs_list": [ 00:33:27.024 { 00:33:27.024 "name": "spare", 00:33:27.024 "uuid": "dcc4ea57-1a06-5d7b-9fab-2c0c863ff1b2", 00:33:27.024 "is_configured": true, 00:33:27.024 "data_offset": 256, 00:33:27.024 "data_size": 7936 00:33:27.024 }, 00:33:27.024 { 00:33:27.024 "name": "BaseBdev2", 00:33:27.024 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:27.024 "is_configured": true, 00:33:27.024 "data_offset": 256, 00:33:27.024 "data_size": 7936 00:33:27.024 } 00:33:27.024 ] 00:33:27.024 }' 00:33:27.024 20:32:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:27.024 20:32:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:27.024 20:32:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:27.024 20:32:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:27.024 20:32:22 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:33:27.592 [2024-10-01 20:32:22.764156] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:33:27.592 [2024-10-01 20:32:22.764557] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:33:27.592 [2024-10-01 20:32:22.764789] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:28.161 "name": "raid_bdev1", 00:33:28.161 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:28.161 "strip_size_kb": 0, 00:33:28.161 "state": "online", 00:33:28.161 "raid_level": "raid1", 00:33:28.161 "superblock": true, 00:33:28.161 "num_base_bdevs": 2, 00:33:28.161 "num_base_bdevs_discovered": 2, 00:33:28.161 "num_base_bdevs_operational": 2, 00:33:28.161 "base_bdevs_list": [ 00:33:28.161 { 00:33:28.161 "name": "spare", 00:33:28.161 "uuid": "dcc4ea57-1a06-5d7b-9fab-2c0c863ff1b2", 00:33:28.161 "is_configured": true, 00:33:28.161 "data_offset": 256, 00:33:28.161 "data_size": 7936 00:33:28.161 }, 00:33:28.161 { 00:33:28.161 "name": "BaseBdev2", 00:33:28.161 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:28.161 "is_configured": true, 00:33:28.161 "data_offset": 256, 00:33:28.161 "data_size": 7936 00:33:28.161 } 00:33:28.161 ] 00:33:28.161 }' 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@709 -- # break 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:28.161 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:28.161 "name": "raid_bdev1", 00:33:28.161 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:28.161 "strip_size_kb": 0, 00:33:28.161 "state": "online", 00:33:28.161 "raid_level": "raid1", 00:33:28.161 "superblock": true, 00:33:28.161 "num_base_bdevs": 2, 00:33:28.161 "num_base_bdevs_discovered": 2, 00:33:28.161 "num_base_bdevs_operational": 2, 00:33:28.161 "base_bdevs_list": [ 00:33:28.161 { 00:33:28.161 "name": "spare", 00:33:28.161 "uuid": "dcc4ea57-1a06-5d7b-9fab-2c0c863ff1b2", 00:33:28.161 "is_configured": true, 00:33:28.161 "data_offset": 256, 00:33:28.161 "data_size": 7936 00:33:28.161 }, 00:33:28.162 { 00:33:28.162 "name": "BaseBdev2", 00:33:28.162 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:28.162 "is_configured": true, 00:33:28.162 "data_offset": 256, 00:33:28.162 "data_size": 7936 00:33:28.162 } 00:33:28.162 ] 00:33:28.162 }' 00:33:28.162 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:28.420 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:33:28.420 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:28.420 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:33:28.420 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:33:28.420 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:28.420 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:28.420 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:28.420 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:28.420 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:28.420 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:28.420 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:28.420 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:28.420 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:28.420 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:28.420 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:28.421 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:28.421 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:28.421 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:28.421 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:28.421 "name": "raid_bdev1", 00:33:28.421 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:28.421 "strip_size_kb": 0, 00:33:28.421 "state": "online", 00:33:28.421 "raid_level": "raid1", 00:33:28.421 "superblock": true, 00:33:28.421 "num_base_bdevs": 2, 00:33:28.421 "num_base_bdevs_discovered": 2, 00:33:28.421 "num_base_bdevs_operational": 2, 00:33:28.421 "base_bdevs_list": [ 00:33:28.421 { 00:33:28.421 "name": "spare", 00:33:28.421 "uuid": "dcc4ea57-1a06-5d7b-9fab-2c0c863ff1b2", 00:33:28.421 "is_configured": true, 00:33:28.421 "data_offset": 256, 00:33:28.421 "data_size": 7936 00:33:28.421 }, 00:33:28.421 { 00:33:28.421 "name": "BaseBdev2", 00:33:28.421 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:28.421 "is_configured": true, 00:33:28.421 "data_offset": 256, 00:33:28.421 "data_size": 7936 00:33:28.421 } 00:33:28.421 ] 00:33:28.421 }' 00:33:28.421 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:28.421 20:32:23 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:28.990 [2024-10-01 20:32:24.011601] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:33:28.990 [2024-10-01 20:32:24.011670] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:33:28.990 [2024-10-01 20:32:24.011882] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:28.990 [2024-10-01 20:32:24.011984] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:28.990 [2024-10-01 20:32:24.012003] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007780 name raid_bdev1, state offline 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # jq length 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:28.990 [2024-10-01 20:32:24.087606] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:33:28.990 [2024-10-01 20:32:24.087689] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:28.990 [2024-10-01 20:32:24.087741] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:33:28.990 [2024-10-01 20:32:24.087760] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:28.990 [2024-10-01 20:32:24.090865] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:28.990 [2024-10-01 20:32:24.090927] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:33:28.990 [2024-10-01 20:32:24.091014] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:33:28.990 [2024-10-01 20:32:24.091087] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:33:28.990 [2024-10-01 20:32:24.091249] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:33:28.990 spare 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:28.990 [2024-10-01 20:32:24.191393] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000007b00 00:33:28.990 [2024-10-01 20:32:24.191477] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:33:28.990 [2024-10-01 20:32:24.191659] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006080 00:33:28.990 [2024-10-01 20:32:24.191900] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000007b00 00:33:28.990 [2024-10-01 20:32:24.191918] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000007b00 00:33:28.990 [2024-10-01 20:32:24.192091] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:28.990 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:29.249 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:29.249 "name": "raid_bdev1", 00:33:29.249 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:29.249 "strip_size_kb": 0, 00:33:29.249 "state": "online", 00:33:29.250 "raid_level": "raid1", 00:33:29.250 "superblock": true, 00:33:29.250 "num_base_bdevs": 2, 00:33:29.250 "num_base_bdevs_discovered": 2, 00:33:29.250 "num_base_bdevs_operational": 2, 00:33:29.250 "base_bdevs_list": [ 00:33:29.250 { 00:33:29.250 "name": "spare", 00:33:29.250 "uuid": "dcc4ea57-1a06-5d7b-9fab-2c0c863ff1b2", 00:33:29.250 "is_configured": true, 00:33:29.250 "data_offset": 256, 00:33:29.250 "data_size": 7936 00:33:29.250 }, 00:33:29.250 { 00:33:29.250 "name": "BaseBdev2", 00:33:29.250 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:29.250 "is_configured": true, 00:33:29.250 "data_offset": 256, 00:33:29.250 "data_size": 7936 00:33:29.250 } 00:33:29.250 ] 00:33:29.250 }' 00:33:29.250 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:29.250 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:29.508 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:33:29.508 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:29.508 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:33:29.508 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:33:29.508 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:29.508 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:29.508 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:29.508 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:29.508 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:29.508 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:29.766 "name": "raid_bdev1", 00:33:29.766 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:29.766 "strip_size_kb": 0, 00:33:29.766 "state": "online", 00:33:29.766 "raid_level": "raid1", 00:33:29.766 "superblock": true, 00:33:29.766 "num_base_bdevs": 2, 00:33:29.766 "num_base_bdevs_discovered": 2, 00:33:29.766 "num_base_bdevs_operational": 2, 00:33:29.766 "base_bdevs_list": [ 00:33:29.766 { 00:33:29.766 "name": "spare", 00:33:29.766 "uuid": "dcc4ea57-1a06-5d7b-9fab-2c0c863ff1b2", 00:33:29.766 "is_configured": true, 00:33:29.766 "data_offset": 256, 00:33:29.766 "data_size": 7936 00:33:29.766 }, 00:33:29.766 { 00:33:29.766 "name": "BaseBdev2", 00:33:29.766 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:29.766 "is_configured": true, 00:33:29.766 "data_offset": 256, 00:33:29.766 "data_size": 7936 00:33:29.766 } 00:33:29.766 ] 00:33:29.766 }' 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:29.766 [2024-10-01 20:32:24.952410] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:29.766 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:29.767 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:29.767 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:29.767 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:29.767 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:29.767 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:29.767 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:29.767 20:32:24 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:29.767 20:32:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:29.767 "name": "raid_bdev1", 00:33:29.767 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:29.767 "strip_size_kb": 0, 00:33:29.767 "state": "online", 00:33:29.767 "raid_level": "raid1", 00:33:29.767 "superblock": true, 00:33:29.767 "num_base_bdevs": 2, 00:33:29.767 "num_base_bdevs_discovered": 1, 00:33:29.767 "num_base_bdevs_operational": 1, 00:33:29.767 "base_bdevs_list": [ 00:33:29.767 { 00:33:29.767 "name": null, 00:33:29.767 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:29.767 "is_configured": false, 00:33:29.767 "data_offset": 0, 00:33:29.767 "data_size": 7936 00:33:29.767 }, 00:33:29.767 { 00:33:29.767 "name": "BaseBdev2", 00:33:29.767 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:29.767 "is_configured": true, 00:33:29.767 "data_offset": 256, 00:33:29.767 "data_size": 7936 00:33:29.767 } 00:33:29.767 ] 00:33:29.767 }' 00:33:29.767 20:32:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:29.767 20:32:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:30.350 20:32:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:33:30.350 20:32:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:30.350 20:32:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:30.350 [2024-10-01 20:32:25.464686] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:33:30.350 [2024-10-01 20:32:25.465012] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:33:30.350 [2024-10-01 20:32:25.465048] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:33:30.350 [2024-10-01 20:32:25.465102] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:33:30.350 [2024-10-01 20:32:25.480416] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006150 00:33:30.350 20:32:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:30.350 20:32:25 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@757 -- # sleep 1 00:33:30.350 [2024-10-01 20:32:25.483298] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:33:31.285 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:31.285 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:31.285 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:31.285 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:31.285 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:31.286 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:31.286 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:31.286 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:31.286 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:31.286 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:31.544 "name": "raid_bdev1", 00:33:31.544 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:31.544 "strip_size_kb": 0, 00:33:31.544 "state": "online", 00:33:31.544 "raid_level": "raid1", 00:33:31.544 "superblock": true, 00:33:31.544 "num_base_bdevs": 2, 00:33:31.544 "num_base_bdevs_discovered": 2, 00:33:31.544 "num_base_bdevs_operational": 2, 00:33:31.544 "process": { 00:33:31.544 "type": "rebuild", 00:33:31.544 "target": "spare", 00:33:31.544 "progress": { 00:33:31.544 "blocks": 2560, 00:33:31.544 "percent": 32 00:33:31.544 } 00:33:31.544 }, 00:33:31.544 "base_bdevs_list": [ 00:33:31.544 { 00:33:31.544 "name": "spare", 00:33:31.544 "uuid": "dcc4ea57-1a06-5d7b-9fab-2c0c863ff1b2", 00:33:31.544 "is_configured": true, 00:33:31.544 "data_offset": 256, 00:33:31.544 "data_size": 7936 00:33:31.544 }, 00:33:31.544 { 00:33:31.544 "name": "BaseBdev2", 00:33:31.544 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:31.544 "is_configured": true, 00:33:31.544 "data_offset": 256, 00:33:31.544 "data_size": 7936 00:33:31.544 } 00:33:31.544 ] 00:33:31.544 }' 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:31.544 [2024-10-01 20:32:26.645524] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:33:31.544 [2024-10-01 20:32:26.695141] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:33:31.544 [2024-10-01 20:32:26.695287] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:31.544 [2024-10-01 20:32:26.695319] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:33:31.544 [2024-10-01 20:32:26.695343] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:31.544 "name": "raid_bdev1", 00:33:31.544 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:31.544 "strip_size_kb": 0, 00:33:31.544 "state": "online", 00:33:31.544 "raid_level": "raid1", 00:33:31.544 "superblock": true, 00:33:31.544 "num_base_bdevs": 2, 00:33:31.544 "num_base_bdevs_discovered": 1, 00:33:31.544 "num_base_bdevs_operational": 1, 00:33:31.544 "base_bdevs_list": [ 00:33:31.544 { 00:33:31.544 "name": null, 00:33:31.544 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:31.544 "is_configured": false, 00:33:31.544 "data_offset": 0, 00:33:31.544 "data_size": 7936 00:33:31.544 }, 00:33:31.544 { 00:33:31.544 "name": "BaseBdev2", 00:33:31.544 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:31.544 "is_configured": true, 00:33:31.544 "data_offset": 256, 00:33:31.544 "data_size": 7936 00:33:31.544 } 00:33:31.544 ] 00:33:31.544 }' 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:31.544 20:32:26 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:32.111 20:32:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:33:32.111 20:32:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:32.111 20:32:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:32.111 [2024-10-01 20:32:27.252092] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:33:32.111 [2024-10-01 20:32:27.252330] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:32.111 [2024-10-01 20:32:27.252418] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:33:32.111 [2024-10-01 20:32:27.252557] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:32.111 [2024-10-01 20:32:27.252935] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:32.111 [2024-10-01 20:32:27.253126] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:33:32.111 [2024-10-01 20:32:27.253346] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:33:32.111 [2024-10-01 20:32:27.253384] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:33:32.111 [2024-10-01 20:32:27.253405] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:33:32.111 [2024-10-01 20:32:27.253475] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:33:32.111 [2024-10-01 20:32:27.269344] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000006220 00:33:32.111 spare 00:33:32.111 20:32:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:32.111 20:32:27 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@764 -- # sleep 1 00:33:32.111 [2024-10-01 20:32:27.272360] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:33:33.048 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:33:33.048 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:33.048 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:33:33.048 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:33:33.049 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:33.049 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:33.049 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:33.049 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:33.049 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:33.049 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:33.307 "name": "raid_bdev1", 00:33:33.307 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:33.307 "strip_size_kb": 0, 00:33:33.307 "state": "online", 00:33:33.307 "raid_level": "raid1", 00:33:33.307 "superblock": true, 00:33:33.307 "num_base_bdevs": 2, 00:33:33.307 "num_base_bdevs_discovered": 2, 00:33:33.307 "num_base_bdevs_operational": 2, 00:33:33.307 "process": { 00:33:33.307 "type": "rebuild", 00:33:33.307 "target": "spare", 00:33:33.307 "progress": { 00:33:33.307 "blocks": 2560, 00:33:33.307 "percent": 32 00:33:33.307 } 00:33:33.307 }, 00:33:33.307 "base_bdevs_list": [ 00:33:33.307 { 00:33:33.307 "name": "spare", 00:33:33.307 "uuid": "dcc4ea57-1a06-5d7b-9fab-2c0c863ff1b2", 00:33:33.307 "is_configured": true, 00:33:33.307 "data_offset": 256, 00:33:33.307 "data_size": 7936 00:33:33.307 }, 00:33:33.307 { 00:33:33.307 "name": "BaseBdev2", 00:33:33.307 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:33.307 "is_configured": true, 00:33:33.307 "data_offset": 256, 00:33:33.307 "data_size": 7936 00:33:33.307 } 00:33:33.307 ] 00:33:33.307 }' 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:33.307 [2024-10-01 20:32:28.443084] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:33:33.307 [2024-10-01 20:32:28.483055] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:33:33.307 [2024-10-01 20:32:28.483248] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:33:33.307 [2024-10-01 20:32:28.483291] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:33:33.307 [2024-10-01 20:32:28.483311] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:33.307 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:33.565 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:33.565 "name": "raid_bdev1", 00:33:33.565 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:33.565 "strip_size_kb": 0, 00:33:33.565 "state": "online", 00:33:33.565 "raid_level": "raid1", 00:33:33.565 "superblock": true, 00:33:33.565 "num_base_bdevs": 2, 00:33:33.565 "num_base_bdevs_discovered": 1, 00:33:33.565 "num_base_bdevs_operational": 1, 00:33:33.565 "base_bdevs_list": [ 00:33:33.565 { 00:33:33.565 "name": null, 00:33:33.565 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:33.565 "is_configured": false, 00:33:33.565 "data_offset": 0, 00:33:33.565 "data_size": 7936 00:33:33.565 }, 00:33:33.565 { 00:33:33.565 "name": "BaseBdev2", 00:33:33.565 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:33.565 "is_configured": true, 00:33:33.565 "data_offset": 256, 00:33:33.565 "data_size": 7936 00:33:33.565 } 00:33:33.565 ] 00:33:33.565 }' 00:33:33.565 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:33.565 20:32:28 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:34.130 "name": "raid_bdev1", 00:33:34.130 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:34.130 "strip_size_kb": 0, 00:33:34.130 "state": "online", 00:33:34.130 "raid_level": "raid1", 00:33:34.130 "superblock": true, 00:33:34.130 "num_base_bdevs": 2, 00:33:34.130 "num_base_bdevs_discovered": 1, 00:33:34.130 "num_base_bdevs_operational": 1, 00:33:34.130 "base_bdevs_list": [ 00:33:34.130 { 00:33:34.130 "name": null, 00:33:34.130 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:34.130 "is_configured": false, 00:33:34.130 "data_offset": 0, 00:33:34.130 "data_size": 7936 00:33:34.130 }, 00:33:34.130 { 00:33:34.130 "name": "BaseBdev2", 00:33:34.130 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:34.130 "is_configured": true, 00:33:34.130 "data_offset": 256, 00:33:34.130 "data_size": 7936 00:33:34.130 } 00:33:34.130 ] 00:33:34.130 }' 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:34.130 [2024-10-01 20:32:29.271612] vbdev_passthru.c: 687:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:33:34.130 [2024-10-01 20:32:29.271918] vbdev_passthru.c: 715:vbdev_passthru_register: *NOTICE*: base bdev opened 00:33:34.130 [2024-10-01 20:32:29.271975] vbdev_passthru.c: 762:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:33:34.130 [2024-10-01 20:32:29.271993] vbdev_passthru.c: 777:vbdev_passthru_register: *NOTICE*: bdev claimed 00:33:34.130 [2024-10-01 20:32:29.272269] vbdev_passthru.c: 790:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:33:34.130 [2024-10-01 20:32:29.272292] vbdev_passthru.c: 791:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:33:34.130 [2024-10-01 20:32:29.272370] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:33:34.130 [2024-10-01 20:32:29.272392] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:33:34.130 [2024-10-01 20:32:29.272406] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:33:34.130 [2024-10-01 20:32:29.272421] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:33:34.130 BaseBdev1 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:34.130 20:32:29 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@775 -- # sleep 1 00:33:35.064 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:33:35.064 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:35.064 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:35.064 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:35.064 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:35.064 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:33:35.064 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:35.064 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:35.064 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:35.064 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:35.064 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:35.064 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:35.064 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:35.064 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:35.064 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:35.322 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:35.322 "name": "raid_bdev1", 00:33:35.322 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:35.322 "strip_size_kb": 0, 00:33:35.322 "state": "online", 00:33:35.322 "raid_level": "raid1", 00:33:35.322 "superblock": true, 00:33:35.322 "num_base_bdevs": 2, 00:33:35.322 "num_base_bdevs_discovered": 1, 00:33:35.322 "num_base_bdevs_operational": 1, 00:33:35.322 "base_bdevs_list": [ 00:33:35.322 { 00:33:35.322 "name": null, 00:33:35.322 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:35.322 "is_configured": false, 00:33:35.322 "data_offset": 0, 00:33:35.322 "data_size": 7936 00:33:35.322 }, 00:33:35.322 { 00:33:35.322 "name": "BaseBdev2", 00:33:35.322 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:35.322 "is_configured": true, 00:33:35.322 "data_offset": 256, 00:33:35.322 "data_size": 7936 00:33:35.322 } 00:33:35.322 ] 00:33:35.322 }' 00:33:35.322 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:35.322 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:35.580 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:33:35.580 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:35.580 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:33:35.580 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:33:35.580 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:35.580 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:35.580 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:35.580 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:35.580 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:35.580 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:35.840 "name": "raid_bdev1", 00:33:35.840 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:35.840 "strip_size_kb": 0, 00:33:35.840 "state": "online", 00:33:35.840 "raid_level": "raid1", 00:33:35.840 "superblock": true, 00:33:35.840 "num_base_bdevs": 2, 00:33:35.840 "num_base_bdevs_discovered": 1, 00:33:35.840 "num_base_bdevs_operational": 1, 00:33:35.840 "base_bdevs_list": [ 00:33:35.840 { 00:33:35.840 "name": null, 00:33:35.840 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:35.840 "is_configured": false, 00:33:35.840 "data_offset": 0, 00:33:35.840 "data_size": 7936 00:33:35.840 }, 00:33:35.840 { 00:33:35.840 "name": "BaseBdev2", 00:33:35.840 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:35.840 "is_configured": true, 00:33:35.840 "data_offset": 256, 00:33:35.840 "data_size": 7936 00:33:35.840 } 00:33:35.840 ] 00:33:35.840 }' 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@650 -- # local es=0 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:35.840 [2024-10-01 20:32:30.964407] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:33:35.840 [2024-10-01 20:32:30.964707] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:33:35.840 [2024-10-01 20:32:30.964765] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:33:35.840 request: 00:33:35.840 { 00:33:35.840 "base_bdev": "BaseBdev1", 00:33:35.840 "raid_bdev": "raid_bdev1", 00:33:35.840 "method": "bdev_raid_add_base_bdev", 00:33:35.840 "req_id": 1 00:33:35.840 } 00:33:35.840 Got JSON-RPC error response 00:33:35.840 response: 00:33:35.840 { 00:33:35.840 "code": -22, 00:33:35.840 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:33:35.840 } 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@653 -- # es=1 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:33:35.840 20:32:30 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@779 -- # sleep 1 00:33:36.776 20:32:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:33:36.776 20:32:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:33:36.776 20:32:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:33:36.776 20:32:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:33:36.776 20:32:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:33:36.776 20:32:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:33:36.776 20:32:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:33:36.776 20:32:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:33:36.776 20:32:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:33:36.776 20:32:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:33:36.776 20:32:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:36.776 20:32:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:36.776 20:32:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:36.776 20:32:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:36.776 20:32:31 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:37.044 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:33:37.044 "name": "raid_bdev1", 00:33:37.044 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:37.044 "strip_size_kb": 0, 00:33:37.044 "state": "online", 00:33:37.044 "raid_level": "raid1", 00:33:37.044 "superblock": true, 00:33:37.044 "num_base_bdevs": 2, 00:33:37.044 "num_base_bdevs_discovered": 1, 00:33:37.044 "num_base_bdevs_operational": 1, 00:33:37.044 "base_bdevs_list": [ 00:33:37.044 { 00:33:37.044 "name": null, 00:33:37.044 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:37.044 "is_configured": false, 00:33:37.044 "data_offset": 0, 00:33:37.044 "data_size": 7936 00:33:37.044 }, 00:33:37.044 { 00:33:37.044 "name": "BaseBdev2", 00:33:37.044 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:37.044 "is_configured": true, 00:33:37.044 "data_offset": 256, 00:33:37.044 "data_size": 7936 00:33:37.044 } 00:33:37.044 ] 00:33:37.044 }' 00:33:37.044 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:33:37.044 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:37.335 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:33:37.335 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:33:37.335 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:33:37.335 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:33:37.335 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:33:37.335 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:33:37.335 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:33:37.335 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:37.335 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:37.335 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:37.335 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:33:37.335 "name": "raid_bdev1", 00:33:37.335 "uuid": "5a82064c-9878-40b0-a7f7-c91ccf055aac", 00:33:37.335 "strip_size_kb": 0, 00:33:37.335 "state": "online", 00:33:37.335 "raid_level": "raid1", 00:33:37.335 "superblock": true, 00:33:37.335 "num_base_bdevs": 2, 00:33:37.335 "num_base_bdevs_discovered": 1, 00:33:37.335 "num_base_bdevs_operational": 1, 00:33:37.335 "base_bdevs_list": [ 00:33:37.335 { 00:33:37.335 "name": null, 00:33:37.335 "uuid": "00000000-0000-0000-0000-000000000000", 00:33:37.335 "is_configured": false, 00:33:37.335 "data_offset": 0, 00:33:37.335 "data_size": 7936 00:33:37.335 }, 00:33:37.335 { 00:33:37.335 "name": "BaseBdev2", 00:33:37.335 "uuid": "d2e4d3da-4702-5e06-8039-f71b9d42f700", 00:33:37.335 "is_configured": true, 00:33:37.335 "data_offset": 256, 00:33:37.335 "data_size": 7936 00:33:37.335 } 00:33:37.335 ] 00:33:37.335 }' 00:33:37.335 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:33:37.593 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:33:37.593 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:33:37.593 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:33:37.593 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@784 -- # killprocess 90583 00:33:37.593 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@950 -- # '[' -z 90583 ']' 00:33:37.593 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # kill -0 90583 00:33:37.593 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@955 -- # uname 00:33:37.593 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:33:37.593 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 90583 00:33:37.593 killing process with pid 90583 00:33:37.593 Received shutdown signal, test time was about 60.000000 seconds 00:33:37.593 00:33:37.593 Latency(us) 00:33:37.593 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:33:37.593 =================================================================================================================== 00:33:37.593 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:33:37.593 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:33:37.593 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:33:37.594 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@968 -- # echo 'killing process with pid 90583' 00:33:37.594 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@969 -- # kill 90583 00:33:37.594 [2024-10-01 20:32:32.685756] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:33:37.594 20:32:32 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@974 -- # wait 90583 00:33:37.594 [2024-10-01 20:32:32.685954] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:33:37.594 [2024-10-01 20:32:32.686040] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:33:37.594 [2024-10-01 20:32:32.686077] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000007b00 name raid_bdev1, state offline 00:33:37.853 [2024-10-01 20:32:32.922857] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:33:39.760 20:32:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@786 -- # return 0 00:33:39.760 00:33:39.760 real 0m19.559s 00:33:39.760 user 0m26.296s 00:33:39.760 sys 0m1.667s 00:33:39.760 ************************************ 00:33:39.760 END TEST raid_rebuild_test_sb_md_interleaved 00:33:39.760 ************************************ 00:33:39.760 20:32:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1126 -- # xtrace_disable 00:33:39.760 20:32:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:33:39.760 20:32:34 bdev_raid -- bdev/bdev_raid.sh@1015 -- # trap - EXIT 00:33:39.760 20:32:34 bdev_raid -- bdev/bdev_raid.sh@1016 -- # cleanup 00:33:39.760 20:32:34 bdev_raid -- bdev/bdev_raid.sh@56 -- # '[' -n 90583 ']' 00:33:39.760 20:32:34 bdev_raid -- bdev/bdev_raid.sh@56 -- # ps -p 90583 00:33:39.760 20:32:34 bdev_raid -- bdev/bdev_raid.sh@60 -- # rm -rf /raidtest 00:33:39.760 00:33:39.760 real 14m23.376s 00:33:39.760 user 19m36.617s 00:33:39.760 sys 2m0.820s 00:33:39.760 20:32:34 bdev_raid -- common/autotest_common.sh@1126 -- # xtrace_disable 00:33:39.760 20:32:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:33:39.760 ************************************ 00:33:39.760 END TEST bdev_raid 00:33:39.760 ************************************ 00:33:39.760 20:32:34 -- spdk/autotest.sh@190 -- # run_test spdkcli_raid /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:33:39.760 20:32:34 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:33:39.760 20:32:34 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:33:39.760 20:32:34 -- common/autotest_common.sh@10 -- # set +x 00:33:39.760 ************************************ 00:33:39.760 START TEST spdkcli_raid 00:33:39.760 ************************************ 00:33:39.760 20:32:34 spdkcli_raid -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:33:39.760 * Looking for test storage... 00:33:39.760 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:33:39.760 20:32:34 spdkcli_raid -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:33:39.760 20:32:34 spdkcli_raid -- common/autotest_common.sh@1681 -- # lcov --version 00:33:39.760 20:32:34 spdkcli_raid -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:33:39.760 20:32:34 spdkcli_raid -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@336 -- # IFS=.-: 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@336 -- # read -ra ver1 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@337 -- # IFS=.-: 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@337 -- # read -ra ver2 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@338 -- # local 'op=<' 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@340 -- # ver1_l=2 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@341 -- # ver2_l=1 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@344 -- # case "$op" in 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@345 -- # : 1 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@365 -- # decimal 1 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@353 -- # local d=1 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@355 -- # echo 1 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@366 -- # decimal 2 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@353 -- # local d=2 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@355 -- # echo 2 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:33:39.760 20:32:34 spdkcli_raid -- scripts/common.sh@368 -- # return 0 00:33:39.760 20:32:34 spdkcli_raid -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:33:39.760 20:32:34 spdkcli_raid -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:33:39.760 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:39.760 --rc genhtml_branch_coverage=1 00:33:39.760 --rc genhtml_function_coverage=1 00:33:39.760 --rc genhtml_legend=1 00:33:39.760 --rc geninfo_all_blocks=1 00:33:39.760 --rc geninfo_unexecuted_blocks=1 00:33:39.760 00:33:39.760 ' 00:33:39.760 20:32:34 spdkcli_raid -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:33:39.760 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:39.760 --rc genhtml_branch_coverage=1 00:33:39.760 --rc genhtml_function_coverage=1 00:33:39.760 --rc genhtml_legend=1 00:33:39.760 --rc geninfo_all_blocks=1 00:33:39.760 --rc geninfo_unexecuted_blocks=1 00:33:39.760 00:33:39.760 ' 00:33:39.760 20:32:34 spdkcli_raid -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:33:39.760 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:39.760 --rc genhtml_branch_coverage=1 00:33:39.760 --rc genhtml_function_coverage=1 00:33:39.760 --rc genhtml_legend=1 00:33:39.760 --rc geninfo_all_blocks=1 00:33:39.760 --rc geninfo_unexecuted_blocks=1 00:33:39.760 00:33:39.760 ' 00:33:39.760 20:32:34 spdkcli_raid -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:33:39.760 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:39.760 --rc genhtml_branch_coverage=1 00:33:39.760 --rc genhtml_function_coverage=1 00:33:39.760 --rc genhtml_legend=1 00:33:39.760 --rc geninfo_all_blocks=1 00:33:39.760 --rc geninfo_unexecuted_blocks=1 00:33:39.761 00:33:39.761 ' 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/raid.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/raid.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/iscsi_tgt/common.sh 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@9 -- # ISCSI_BRIDGE=iscsi_br 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@10 -- # INITIATOR_INTERFACE=spdk_init_int 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@11 -- # INITIATOR_BRIDGE=init_br 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@12 -- # TARGET_NAMESPACE=spdk_iscsi_ns 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@13 -- # TARGET_NS_CMD=(ip netns exec "$TARGET_NAMESPACE") 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@14 -- # TARGET_INTERFACE=spdk_tgt_int 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@15 -- # TARGET_INTERFACE2=spdk_tgt_int2 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@16 -- # TARGET_BRIDGE=tgt_br 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@17 -- # TARGET_BRIDGE2=tgt_br2 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@20 -- # TARGET_IP=10.0.0.1 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@21 -- # TARGET_IP2=10.0.0.3 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@22 -- # INITIATOR_IP=10.0.0.2 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@23 -- # ISCSI_PORT=3260 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@24 -- # NETMASK=10.0.0.2/32 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@25 -- # INITIATOR_TAG=2 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@26 -- # INITIATOR_NAME=ANY 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@27 -- # PORTAL_TAG=1 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@28 -- # ISCSI_APP=("${TARGET_NS_CMD[@]}" "${ISCSI_APP[@]}") 00:33:39.761 20:32:34 spdkcli_raid -- iscsi_tgt/common.sh@29 -- # ISCSI_TEST_CORE_MASK=0xF 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/raid.sh@12 -- # MATCH_FILE=spdkcli_raid.test 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/raid.sh@13 -- # SPDKCLI_BRANCH=/bdevs 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/raid.sh@14 -- # dirname /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/raid.sh@14 -- # readlink -f /home/vagrant/spdk_repo/spdk/test/spdkcli 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/raid.sh@14 -- # testdir=/home/vagrant/spdk_repo/spdk/test/spdkcli 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/raid.sh@15 -- # . /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/raid.sh@17 -- # trap cleanup EXIT 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/raid.sh@19 -- # timing_enter run_spdk_tgt 00:33:39.761 20:32:34 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:33:39.761 20:32:34 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:33:39.761 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/raid.sh@20 -- # run_spdk_tgt 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/common.sh@27 -- # spdk_tgt_pid=91271 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/common.sh@26 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:33:39.761 20:32:34 spdkcli_raid -- spdkcli/common.sh@28 -- # waitforlisten 91271 00:33:39.761 20:32:34 spdkcli_raid -- common/autotest_common.sh@831 -- # '[' -z 91271 ']' 00:33:39.761 20:32:34 spdkcli_raid -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:39.761 20:32:34 spdkcli_raid -- common/autotest_common.sh@836 -- # local max_retries=100 00:33:39.761 20:32:34 spdkcli_raid -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:39.761 20:32:34 spdkcli_raid -- common/autotest_common.sh@840 -- # xtrace_disable 00:33:39.761 20:32:34 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:33:40.020 [2024-10-01 20:32:35.089128] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:33:40.020 [2024-10-01 20:32:35.089628] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid91271 ] 00:33:40.020 [2024-10-01 20:32:35.260110] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:33:40.283 [2024-10-01 20:32:35.466895] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:33:40.283 [2024-10-01 20:32:35.466908] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:33:41.660 20:32:36 spdkcli_raid -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:33:41.660 20:32:36 spdkcli_raid -- common/autotest_common.sh@864 -- # return 0 00:33:41.660 20:32:36 spdkcli_raid -- spdkcli/raid.sh@21 -- # timing_exit run_spdk_tgt 00:33:41.660 20:32:36 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:33:41.660 20:32:36 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:33:41.660 20:32:36 spdkcli_raid -- spdkcli/raid.sh@23 -- # timing_enter spdkcli_create_malloc 00:33:41.660 20:32:36 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:33:41.660 20:32:36 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:33:41.660 20:32:36 spdkcli_raid -- spdkcli/raid.sh@26 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 8 512 Malloc1'\'' '\''Malloc1'\'' True 00:33:41.660 '\''/bdevs/malloc create 8 512 Malloc2'\'' '\''Malloc2'\'' True 00:33:41.660 ' 00:33:43.034 Executing command: ['/bdevs/malloc create 8 512 Malloc1', 'Malloc1', True] 00:33:43.034 Executing command: ['/bdevs/malloc create 8 512 Malloc2', 'Malloc2', True] 00:33:43.034 20:32:38 spdkcli_raid -- spdkcli/raid.sh@27 -- # timing_exit spdkcli_create_malloc 00:33:43.034 20:32:38 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:33:43.034 20:32:38 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:33:43.293 20:32:38 spdkcli_raid -- spdkcli/raid.sh@29 -- # timing_enter spdkcli_create_raid 00:33:43.293 20:32:38 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:33:43.293 20:32:38 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:33:43.293 20:32:38 spdkcli_raid -- spdkcli/raid.sh@31 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4'\'' '\''testraid'\'' True 00:33:43.293 ' 00:33:44.231 Executing command: ['/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4', 'testraid', True] 00:33:44.489 20:32:39 spdkcli_raid -- spdkcli/raid.sh@32 -- # timing_exit spdkcli_create_raid 00:33:44.490 20:32:39 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:33:44.490 20:32:39 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:33:44.490 20:32:39 spdkcli_raid -- spdkcli/raid.sh@34 -- # timing_enter spdkcli_check_match 00:33:44.490 20:32:39 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:33:44.490 20:32:39 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:33:44.490 20:32:39 spdkcli_raid -- spdkcli/raid.sh@35 -- # check_match 00:33:44.490 20:32:39 spdkcli_raid -- spdkcli/common.sh@44 -- # /home/vagrant/spdk_repo/spdk/scripts/spdkcli.py ll /bdevs 00:33:45.058 20:32:40 spdkcli_raid -- spdkcli/common.sh@45 -- # /home/vagrant/spdk_repo/spdk/test/app/match/match /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test.match 00:33:45.058 20:32:40 spdkcli_raid -- spdkcli/common.sh@46 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test 00:33:45.058 20:32:40 spdkcli_raid -- spdkcli/raid.sh@36 -- # timing_exit spdkcli_check_match 00:33:45.058 20:32:40 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:33:45.058 20:32:40 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:33:45.058 20:32:40 spdkcli_raid -- spdkcli/raid.sh@38 -- # timing_enter spdkcli_delete_raid 00:33:45.058 20:32:40 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:33:45.058 20:32:40 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:33:45.058 20:32:40 spdkcli_raid -- spdkcli/raid.sh@40 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume delete testraid'\'' '\'''\'' True 00:33:45.058 ' 00:33:46.437 Executing command: ['/bdevs/raid_volume delete testraid', '', True] 00:33:46.437 20:32:41 spdkcli_raid -- spdkcli/raid.sh@41 -- # timing_exit spdkcli_delete_raid 00:33:46.437 20:32:41 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:33:46.437 20:32:41 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:33:46.437 20:32:41 spdkcli_raid -- spdkcli/raid.sh@43 -- # timing_enter spdkcli_delete_malloc 00:33:46.437 20:32:41 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:33:46.437 20:32:41 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:33:46.437 20:32:41 spdkcli_raid -- spdkcli/raid.sh@46 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc delete Malloc1'\'' '\'''\'' True 00:33:46.437 '\''/bdevs/malloc delete Malloc2'\'' '\'''\'' True 00:33:46.437 ' 00:33:47.815 Executing command: ['/bdevs/malloc delete Malloc1', '', True] 00:33:47.815 Executing command: ['/bdevs/malloc delete Malloc2', '', True] 00:33:47.815 20:32:42 spdkcli_raid -- spdkcli/raid.sh@47 -- # timing_exit spdkcli_delete_malloc 00:33:47.815 20:32:42 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:33:47.815 20:32:42 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:33:47.815 20:32:43 spdkcli_raid -- spdkcli/raid.sh@49 -- # killprocess 91271 00:33:47.815 20:32:43 spdkcli_raid -- common/autotest_common.sh@950 -- # '[' -z 91271 ']' 00:33:47.815 20:32:43 spdkcli_raid -- common/autotest_common.sh@954 -- # kill -0 91271 00:33:47.815 20:32:43 spdkcli_raid -- common/autotest_common.sh@955 -- # uname 00:33:47.815 20:32:43 spdkcli_raid -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:33:47.815 20:32:43 spdkcli_raid -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 91271 00:33:47.815 20:32:43 spdkcli_raid -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:33:47.815 killing process with pid 91271 00:33:47.815 20:32:43 spdkcli_raid -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:33:47.815 20:32:43 spdkcli_raid -- common/autotest_common.sh@968 -- # echo 'killing process with pid 91271' 00:33:47.815 20:32:43 spdkcli_raid -- common/autotest_common.sh@969 -- # kill 91271 00:33:47.815 20:32:43 spdkcli_raid -- common/autotest_common.sh@974 -- # wait 91271 00:33:51.104 20:32:46 spdkcli_raid -- spdkcli/raid.sh@1 -- # cleanup 00:33:51.104 20:32:46 spdkcli_raid -- spdkcli/common.sh@10 -- # '[' -n 91271 ']' 00:33:51.104 20:32:46 spdkcli_raid -- spdkcli/common.sh@11 -- # killprocess 91271 00:33:51.104 20:32:46 spdkcli_raid -- common/autotest_common.sh@950 -- # '[' -z 91271 ']' 00:33:51.104 20:32:46 spdkcli_raid -- common/autotest_common.sh@954 -- # kill -0 91271 00:33:51.104 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 954: kill: (91271) - No such process 00:33:51.104 Process with pid 91271 is not found 00:33:51.104 20:32:46 spdkcli_raid -- common/autotest_common.sh@977 -- # echo 'Process with pid 91271 is not found' 00:33:51.104 20:32:46 spdkcli_raid -- spdkcli/common.sh@13 -- # '[' -n '' ']' 00:33:51.104 20:32:46 spdkcli_raid -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:33:51.104 20:32:46 spdkcli_raid -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:33:51.104 20:32:46 spdkcli_raid -- spdkcli/common.sh@22 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_raid.test /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:33:51.104 ************************************ 00:33:51.104 END TEST spdkcli_raid 00:33:51.104 ************************************ 00:33:51.104 00:33:51.104 real 0m11.524s 00:33:51.104 user 0m23.346s 00:33:51.104 sys 0m1.336s 00:33:51.104 20:32:46 spdkcli_raid -- common/autotest_common.sh@1126 -- # xtrace_disable 00:33:51.104 20:32:46 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:33:51.104 20:32:46 -- spdk/autotest.sh@191 -- # run_test blockdev_raid5f /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:33:51.104 20:32:46 -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:33:51.104 20:32:46 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:33:51.104 20:32:46 -- common/autotest_common.sh@10 -- # set +x 00:33:51.104 ************************************ 00:33:51.104 START TEST blockdev_raid5f 00:33:51.104 ************************************ 00:33:51.104 20:32:46 blockdev_raid5f -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:33:51.364 * Looking for test storage... 00:33:51.364 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:33:51.364 20:32:46 blockdev_raid5f -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:33:51.364 20:32:46 blockdev_raid5f -- common/autotest_common.sh@1681 -- # lcov --version 00:33:51.364 20:32:46 blockdev_raid5f -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:33:51.364 20:32:46 blockdev_raid5f -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@333 -- # local ver1 ver1_l 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@334 -- # local ver2 ver2_l 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@336 -- # IFS=.-: 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@336 -- # read -ra ver1 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@337 -- # IFS=.-: 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@337 -- # read -ra ver2 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@338 -- # local 'op=<' 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@340 -- # ver1_l=2 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@341 -- # ver2_l=1 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@344 -- # case "$op" in 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@345 -- # : 1 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@364 -- # (( v = 0 )) 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@365 -- # decimal 1 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@353 -- # local d=1 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@355 -- # echo 1 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@365 -- # ver1[v]=1 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@366 -- # decimal 2 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@353 -- # local d=2 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@355 -- # echo 2 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@366 -- # ver2[v]=2 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:33:51.364 20:32:46 blockdev_raid5f -- scripts/common.sh@368 -- # return 0 00:33:51.364 20:32:46 blockdev_raid5f -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:33:51.364 20:32:46 blockdev_raid5f -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:33:51.364 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:51.364 --rc genhtml_branch_coverage=1 00:33:51.364 --rc genhtml_function_coverage=1 00:33:51.364 --rc genhtml_legend=1 00:33:51.364 --rc geninfo_all_blocks=1 00:33:51.364 --rc geninfo_unexecuted_blocks=1 00:33:51.364 00:33:51.364 ' 00:33:51.364 20:32:46 blockdev_raid5f -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:33:51.364 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:51.364 --rc genhtml_branch_coverage=1 00:33:51.364 --rc genhtml_function_coverage=1 00:33:51.364 --rc genhtml_legend=1 00:33:51.364 --rc geninfo_all_blocks=1 00:33:51.364 --rc geninfo_unexecuted_blocks=1 00:33:51.364 00:33:51.364 ' 00:33:51.364 20:32:46 blockdev_raid5f -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:33:51.364 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:51.364 --rc genhtml_branch_coverage=1 00:33:51.364 --rc genhtml_function_coverage=1 00:33:51.364 --rc genhtml_legend=1 00:33:51.364 --rc geninfo_all_blocks=1 00:33:51.364 --rc geninfo_unexecuted_blocks=1 00:33:51.364 00:33:51.364 ' 00:33:51.364 20:32:46 blockdev_raid5f -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:33:51.364 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:51.364 --rc genhtml_branch_coverage=1 00:33:51.364 --rc genhtml_function_coverage=1 00:33:51.364 --rc genhtml_legend=1 00:33:51.364 --rc geninfo_all_blocks=1 00:33:51.364 --rc geninfo_unexecuted_blocks=1 00:33:51.364 00:33:51.364 ' 00:33:51.364 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:33:51.364 20:32:46 blockdev_raid5f -- bdev/nbd_common.sh@6 -- # set -e 00:33:51.364 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@12 -- # rpc_py=rpc_cmd 00:33:51.364 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@13 -- # conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:33:51.364 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@14 -- # nonenclosed_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json 00:33:51.364 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@15 -- # nonarray_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@17 -- # export RPC_PIPE_TIMEOUT=30 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@17 -- # RPC_PIPE_TIMEOUT=30 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@20 -- # : 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@671 -- # QOS_DEV_1=Malloc_0 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@672 -- # QOS_DEV_2=Null_1 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@673 -- # QOS_RUN_TIME=5 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@675 -- # uname -s 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@675 -- # '[' Linux = Linux ']' 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@677 -- # PRE_RESERVED_MEM=0 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@683 -- # test_type=raid5f 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@684 -- # crypto_device= 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@685 -- # dek= 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@686 -- # env_ctx= 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@687 -- # wait_for_rpc= 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@688 -- # '[' -n '' ']' 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@691 -- # [[ raid5f == bdev ]] 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@691 -- # [[ raid5f == crypto_* ]] 00:33:51.365 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@694 -- # start_spdk_tgt 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@47 -- # spdk_tgt_pid=91563 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@48 -- # trap 'killprocess "$spdk_tgt_pid"; exit 1' SIGINT SIGTERM EXIT 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@49 -- # waitforlisten 91563 00:33:51.365 20:32:46 blockdev_raid5f -- bdev/blockdev.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt '' '' 00:33:51.365 20:32:46 blockdev_raid5f -- common/autotest_common.sh@831 -- # '[' -z 91563 ']' 00:33:51.365 20:32:46 blockdev_raid5f -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:51.365 20:32:46 blockdev_raid5f -- common/autotest_common.sh@836 -- # local max_retries=100 00:33:51.365 20:32:46 blockdev_raid5f -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:51.365 20:32:46 blockdev_raid5f -- common/autotest_common.sh@840 -- # xtrace_disable 00:33:51.365 20:32:46 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:33:51.623 [2024-10-01 20:32:46.663552] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:33:51.623 [2024-10-01 20:32:46.664065] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid91563 ] 00:33:51.623 [2024-10-01 20:32:46.847410] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:52.190 [2024-10-01 20:32:47.153896] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:33:53.126 20:32:48 blockdev_raid5f -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:33:53.126 20:32:48 blockdev_raid5f -- common/autotest_common.sh@864 -- # return 0 00:33:53.126 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@695 -- # case "$test_type" in 00:33:53.126 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@727 -- # setup_raid5f_conf 00:33:53.126 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@281 -- # rpc_cmd 00:33:53.126 20:32:48 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:53.126 20:32:48 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:33:53.126 Malloc0 00:33:53.385 Malloc1 00:33:53.385 Malloc2 00:33:53.385 20:32:48 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:53.385 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@738 -- # rpc_cmd bdev_wait_for_examine 00:33:53.385 20:32:48 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:53.385 20:32:48 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:33:53.385 20:32:48 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:53.385 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@741 -- # cat 00:33:53.385 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@741 -- # rpc_cmd save_subsystem_config -n accel 00:33:53.385 20:32:48 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:53.385 20:32:48 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:33:53.385 20:32:48 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:53.385 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@741 -- # rpc_cmd save_subsystem_config -n bdev 00:33:53.385 20:32:48 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:53.385 20:32:48 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:33:53.385 20:32:48 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:53.385 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@741 -- # rpc_cmd save_subsystem_config -n iobuf 00:33:53.385 20:32:48 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:53.385 20:32:48 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:33:53.385 20:32:48 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:53.385 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@749 -- # mapfile -t bdevs 00:33:53.385 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@749 -- # rpc_cmd bdev_get_bdevs 00:33:53.385 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@749 -- # jq -r '.[] | select(.claimed == false)' 00:33:53.385 20:32:48 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:53.385 20:32:48 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:33:53.385 20:32:48 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:53.385 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@750 -- # mapfile -t bdevs_name 00:33:53.385 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@750 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "9e952634-cec7-4ee5-b5dd-825034055a51"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "9e952634-cec7-4ee5-b5dd-825034055a51",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "9e952634-cec7-4ee5-b5dd-825034055a51",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "7a906a5c-da82-4570-8225-84806d3225ed",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "29123d65-a63b-44fb-9cc0-358d9ff3185e",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "bab1c060-9a38-421b-b614-13bb7f187036",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:33:53.385 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@750 -- # jq -r .name 00:33:53.655 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@751 -- # bdev_list=("${bdevs_name[@]}") 00:33:53.655 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@753 -- # hello_world_bdev=raid5f 00:33:53.655 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@754 -- # trap - SIGINT SIGTERM EXIT 00:33:53.655 20:32:48 blockdev_raid5f -- bdev/blockdev.sh@755 -- # killprocess 91563 00:33:53.655 20:32:48 blockdev_raid5f -- common/autotest_common.sh@950 -- # '[' -z 91563 ']' 00:33:53.655 20:32:48 blockdev_raid5f -- common/autotest_common.sh@954 -- # kill -0 91563 00:33:53.655 20:32:48 blockdev_raid5f -- common/autotest_common.sh@955 -- # uname 00:33:53.655 20:32:48 blockdev_raid5f -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:33:53.655 20:32:48 blockdev_raid5f -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 91563 00:33:53.655 killing process with pid 91563 00:33:53.655 20:32:48 blockdev_raid5f -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:33:53.655 20:32:48 blockdev_raid5f -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:33:53.655 20:32:48 blockdev_raid5f -- common/autotest_common.sh@968 -- # echo 'killing process with pid 91563' 00:33:53.655 20:32:48 blockdev_raid5f -- common/autotest_common.sh@969 -- # kill 91563 00:33:53.655 20:32:48 blockdev_raid5f -- common/autotest_common.sh@974 -- # wait 91563 00:33:56.975 20:32:52 blockdev_raid5f -- bdev/blockdev.sh@759 -- # trap cleanup SIGINT SIGTERM EXIT 00:33:56.975 20:32:52 blockdev_raid5f -- bdev/blockdev.sh@761 -- # run_test bdev_hello_world /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:33:56.975 20:32:52 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:33:56.975 20:32:52 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:33:56.975 20:32:52 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:33:57.234 ************************************ 00:33:57.234 START TEST bdev_hello_world 00:33:57.234 ************************************ 00:33:57.234 20:32:52 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:33:57.234 [2024-10-01 20:32:52.339601] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:33:57.235 [2024-10-01 20:32:52.339788] [ DPDK EAL parameters: hello_bdev --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid91641 ] 00:33:57.493 [2024-10-01 20:32:52.512533] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:33:57.754 [2024-10-01 20:32:52.785248] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:33:58.692 [2024-10-01 20:32:53.590761] hello_bdev.c: 222:hello_start: *NOTICE*: Successfully started the application 00:33:58.692 [2024-10-01 20:32:53.590840] hello_bdev.c: 231:hello_start: *NOTICE*: Opening the bdev raid5f 00:33:58.692 [2024-10-01 20:32:53.590868] hello_bdev.c: 244:hello_start: *NOTICE*: Opening io channel 00:33:58.692 [2024-10-01 20:32:53.591486] hello_bdev.c: 138:hello_write: *NOTICE*: Writing to the bdev 00:33:58.692 [2024-10-01 20:32:53.591651] hello_bdev.c: 117:write_complete: *NOTICE*: bdev io write completed successfully 00:33:58.692 [2024-10-01 20:32:53.591680] hello_bdev.c: 84:hello_read: *NOTICE*: Reading io 00:33:58.692 [2024-10-01 20:32:53.591782] hello_bdev.c: 65:read_complete: *NOTICE*: Read string from bdev : Hello World! 00:33:58.692 00:33:58.692 [2024-10-01 20:32:53.591821] hello_bdev.c: 74:read_complete: *NOTICE*: Stopping app 00:34:00.598 00:34:00.598 real 0m3.434s 00:34:00.598 user 0m2.877s 00:34:00.598 sys 0m0.414s 00:34:00.598 20:32:55 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:00.598 ************************************ 00:34:00.598 END TEST bdev_hello_world 00:34:00.598 ************************************ 00:34:00.598 20:32:55 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@10 -- # set +x 00:34:00.598 20:32:55 blockdev_raid5f -- bdev/blockdev.sh@762 -- # run_test bdev_bounds bdev_bounds '' 00:34:00.598 20:32:55 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:34:00.598 20:32:55 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:34:00.598 20:32:55 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:34:00.598 ************************************ 00:34:00.598 START TEST bdev_bounds 00:34:00.598 ************************************ 00:34:00.598 Process bdevio pid: 91699 00:34:00.598 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:00.598 20:32:55 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1125 -- # bdev_bounds '' 00:34:00.598 20:32:55 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@291 -- # bdevio_pid=91699 00:34:00.598 20:32:55 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@292 -- # trap 'cleanup; killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT 00:34:00.598 20:32:55 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@293 -- # echo 'Process bdevio pid: 91699' 00:34:00.598 20:32:55 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@290 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/bdevio -w -s 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:34:00.598 20:32:55 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@294 -- # waitforlisten 91699 00:34:00.598 20:32:55 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@831 -- # '[' -z 91699 ']' 00:34:00.598 20:32:55 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:00.598 20:32:55 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@836 -- # local max_retries=100 00:34:00.598 20:32:55 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:00.598 20:32:55 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@840 -- # xtrace_disable 00:34:00.599 20:32:55 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:34:00.599 [2024-10-01 20:32:55.845462] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:34:00.599 [2024-10-01 20:32:55.846068] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 0 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid91699 ] 00:34:00.858 [2024-10-01 20:32:56.033235] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 3 00:34:01.117 [2024-10-01 20:32:56.325503] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:34:01.117 [2024-10-01 20:32:56.325691] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:34:01.117 [2024-10-01 20:32:56.325771] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:34:02.065 20:32:57 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:34:02.065 20:32:57 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@864 -- # return 0 00:34:02.065 20:32:57 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@295 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/tests.py perform_tests 00:34:02.353 I/O targets: 00:34:02.353 raid5f: 131072 blocks of 512 bytes (64 MiB) 00:34:02.353 00:34:02.353 00:34:02.353 CUnit - A unit testing framework for C - Version 2.1-3 00:34:02.353 http://cunit.sourceforge.net/ 00:34:02.353 00:34:02.353 00:34:02.353 Suite: bdevio tests on: raid5f 00:34:02.353 Test: blockdev write read block ...passed 00:34:02.353 Test: blockdev write zeroes read block ...passed 00:34:02.353 Test: blockdev write zeroes read no split ...passed 00:34:02.353 Test: blockdev write zeroes read split ...passed 00:34:02.612 Test: blockdev write zeroes read split partial ...passed 00:34:02.612 Test: blockdev reset ...passed 00:34:02.612 Test: blockdev write read 8 blocks ...passed 00:34:02.612 Test: blockdev write read size > 128k ...passed 00:34:02.612 Test: blockdev write read invalid size ...passed 00:34:02.612 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:34:02.612 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:34:02.612 Test: blockdev write read max offset ...passed 00:34:02.612 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:34:02.612 Test: blockdev writev readv 8 blocks ...passed 00:34:02.612 Test: blockdev writev readv 30 x 1block ...passed 00:34:02.612 Test: blockdev writev readv block ...passed 00:34:02.612 Test: blockdev writev readv size > 128k ...passed 00:34:02.612 Test: blockdev writev readv size > 128k in two iovs ...passed 00:34:02.612 Test: blockdev comparev and writev ...passed 00:34:02.612 Test: blockdev nvme passthru rw ...passed 00:34:02.612 Test: blockdev nvme passthru vendor specific ...passed 00:34:02.612 Test: blockdev nvme admin passthru ...passed 00:34:02.612 Test: blockdev copy ...passed 00:34:02.612 00:34:02.612 Run Summary: Type Total Ran Passed Failed Inactive 00:34:02.612 suites 1 1 n/a 0 0 00:34:02.612 tests 23 23 23 0 0 00:34:02.612 asserts 130 130 130 0 n/a 00:34:02.612 00:34:02.612 Elapsed time = 0.622 seconds 00:34:02.612 0 00:34:02.612 20:32:57 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@296 -- # killprocess 91699 00:34:02.612 20:32:57 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@950 -- # '[' -z 91699 ']' 00:34:02.612 20:32:57 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@954 -- # kill -0 91699 00:34:02.612 20:32:57 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@955 -- # uname 00:34:02.612 20:32:57 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:34:02.612 20:32:57 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 91699 00:34:02.612 20:32:57 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:34:02.612 20:32:57 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:34:02.612 20:32:57 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@968 -- # echo 'killing process with pid 91699' 00:34:02.612 killing process with pid 91699 00:34:02.612 20:32:57 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@969 -- # kill 91699 00:34:02.612 20:32:57 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@974 -- # wait 91699 00:34:05.141 20:32:59 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@297 -- # trap - SIGINT SIGTERM EXIT 00:34:05.141 00:34:05.141 real 0m4.110s 00:34:05.141 user 0m10.054s 00:34:05.141 sys 0m0.593s 00:34:05.141 20:32:59 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:05.141 20:32:59 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:34:05.141 ************************************ 00:34:05.141 END TEST bdev_bounds 00:34:05.141 ************************************ 00:34:05.141 20:32:59 blockdev_raid5f -- bdev/blockdev.sh@763 -- # run_test bdev_nbd nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:34:05.141 20:32:59 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:34:05.141 20:32:59 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:34:05.141 20:32:59 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:34:05.141 ************************************ 00:34:05.141 START TEST bdev_nbd 00:34:05.141 ************************************ 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1125 -- # nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@301 -- # uname -s 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@301 -- # [[ Linux == Linux ]] 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@304 -- # local conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@305 -- # bdev_all=('raid5f') 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@305 -- # local bdev_all 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@306 -- # local bdev_num=1 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # [[ -e /sys/module/nbd ]] 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@312 -- # nbd_all=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@312 -- # local nbd_all 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # bdev_num=1 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@315 -- # nbd_list=('/dev/nbd0') 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@315 -- # local nbd_list 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@316 -- # bdev_list=('raid5f') 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@316 -- # local bdev_list 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@319 -- # nbd_pid=91771 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@318 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-nbd.sock -i 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@320 -- # trap 'cleanup; killprocess $nbd_pid' SIGINT SIGTERM EXIT 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@321 -- # waitforlisten 91771 /var/tmp/spdk-nbd.sock 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@831 -- # '[' -z 91771 ']' 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@836 -- # local max_retries=100 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:34:05.141 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@840 -- # xtrace_disable 00:34:05.141 20:32:59 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:34:05.141 [2024-10-01 20:33:00.014897] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:34:05.141 [2024-10-01 20:33:00.015099] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:34:05.141 [2024-10-01 20:33:00.192504] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:05.399 [2024-10-01 20:33:00.467161] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:34:06.333 20:33:01 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:34:06.333 20:33:01 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@864 -- # return 0 00:34:06.333 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@323 -- # nbd_rpc_start_stop_verify /var/tmp/spdk-nbd.sock raid5f 00:34:06.333 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@113 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:34:06.333 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # bdev_list=('raid5f') 00:34:06.333 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # local bdev_list 00:34:06.333 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@116 -- # nbd_start_disks_without_nbd_idx /var/tmp/spdk-nbd.sock raid5f 00:34:06.333 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@22 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:34:06.333 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # bdev_list=('raid5f') 00:34:06.333 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # local bdev_list 00:34:06.333 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@24 -- # local i 00:34:06.333 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@25 -- # local nbd_device 00:34:06.333 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i = 0 )) 00:34:06.333 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:34:06.333 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd0 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd0 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd0 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@869 -- # local i 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # break 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:34:06.592 1+0 records in 00:34:06.592 1+0 records out 00:34:06.592 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000287795 s, 14.2 MB/s 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # size=4096 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # return 0 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:34:06.592 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:34:06.850 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # nbd_disks_json='[ 00:34:06.850 { 00:34:06.850 "nbd_device": "/dev/nbd0", 00:34:06.850 "bdev_name": "raid5f" 00:34:06.850 } 00:34:06.850 ]' 00:34:06.850 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # nbd_disks_name=($(echo "${nbd_disks_json}" | jq -r '.[] | .nbd_device')) 00:34:06.850 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # echo '[ 00:34:06.850 { 00:34:06.850 "nbd_device": "/dev/nbd0", 00:34:06.850 "bdev_name": "raid5f" 00:34:06.850 } 00:34:06.850 ]' 00:34:06.850 20:33:01 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # jq -r '.[] | .nbd_device' 00:34:06.850 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@120 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:34:06.850 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:34:06.850 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:34:06.850 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:34:06.850 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:34:06.850 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:34:06.850 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:34:07.108 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:34:07.108 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:34:07.108 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:34:07.108 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:34:07.108 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:34:07.108 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:34:07.108 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:34:07.108 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:34:07.108 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:34:07.108 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:34:07.108 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:34:07.366 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:34:07.366 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:34:07.366 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # count=0 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@123 -- # '[' 0 -ne 0 ']' 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@127 -- # return 0 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@324 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # bdev_list=('raid5f') 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # local bdev_list 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0') 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # local nbd_list 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # bdev_list=('raid5f') 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # local bdev_list 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # local nbd_list 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@12 -- # local i 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:34:07.625 20:33:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f /dev/nbd0 00:34:07.883 /dev/nbd0 00:34:07.883 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:34:07.883 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:34:07.883 20:33:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:34:07.883 20:33:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@869 -- # local i 00:34:07.883 20:33:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:34:07.884 20:33:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:34:07.884 20:33:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:34:07.884 20:33:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # break 00:34:07.884 20:33:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:34:07.884 20:33:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:34:07.884 20:33:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:34:07.884 1+0 records in 00:34:07.884 1+0 records out 00:34:07.884 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000280443 s, 14.6 MB/s 00:34:07.884 20:33:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:07.884 20:33:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # size=4096 00:34:07.884 20:33:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:34:07.884 20:33:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:34:07.884 20:33:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # return 0 00:34:07.884 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:34:07.884 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:34:07.884 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:34:07.884 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:34:07.884 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:34:08.142 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:34:08.142 { 00:34:08.142 "nbd_device": "/dev/nbd0", 00:34:08.142 "bdev_name": "raid5f" 00:34:08.142 } 00:34:08.142 ]' 00:34:08.142 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[ 00:34:08.142 { 00:34:08.142 "nbd_device": "/dev/nbd0", 00:34:08.142 "bdev_name": "raid5f" 00:34:08.142 } 00:34:08.142 ]' 00:34:08.142 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=1 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 1 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # count=1 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@96 -- # '[' 1 -ne 1 ']' 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify /dev/nbd0 write 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=write 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest bs=4096 count=256 00:34:08.401 256+0 records in 00:34:08.401 256+0 records out 00:34:08.401 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0107526 s, 97.5 MB/s 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:34:08.401 256+0 records in 00:34:08.401 256+0 records out 00:34:08.401 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0468591 s, 22.4 MB/s 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify /dev/nbd0 verify 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=verify 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest /dev/nbd0 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:34:08.401 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:34:08.660 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:34:08.660 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:34:08.660 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:34:08.660 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:34:08.660 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:34:08.660 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:34:08.660 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:34:08.660 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:34:08.660 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:34:08.660 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:34:08.660 20:33:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:34:08.918 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:34:08.918 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:34:08.918 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:34:09.177 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:34:09.177 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:34:09.177 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:34:09.177 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:34:09.177 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:34:09.177 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:34:09.177 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # count=0 00:34:09.177 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:34:09.177 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@109 -- # return 0 00:34:09.177 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@325 -- # nbd_with_lvol_verify /var/tmp/spdk-nbd.sock /dev/nbd0 00:34:09.177 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@131 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:34:09.177 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@132 -- # local nbd=/dev/nbd0 00:34:09.177 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@134 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create -b malloc_lvol_verify 16 512 00:34:09.471 malloc_lvol_verify 00:34:09.471 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@135 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create_lvstore malloc_lvol_verify lvs 00:34:09.731 99b6dbd5-72cf-477a-b08e-567470fa7fb3 00:34:09.731 20:33:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@136 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create lvol 4 -l lvs 00:34:09.990 742228ff-2c81-4c69-a9af-c29755ca32c4 00:34:09.990 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@137 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk lvs/lvol /dev/nbd0 00:34:10.249 /dev/nbd0 00:34:10.249 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@139 -- # wait_for_nbd_set_capacity /dev/nbd0 00:34:10.249 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@146 -- # local nbd=nbd0 00:34:10.249 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@148 -- # [[ -e /sys/block/nbd0/size ]] 00:34:10.249 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@150 -- # (( 8192 == 0 )) 00:34:10.249 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@141 -- # mkfs.ext4 /dev/nbd0 00:34:10.249 mke2fs 1.47.0 (5-Feb-2023) 00:34:10.249 Discarding device blocks: 0/4096 done 00:34:10.249 Creating filesystem with 4096 1k blocks and 1024 inodes 00:34:10.249 00:34:10.249 Allocating group tables: 0/1 done 00:34:10.249 Writing inode tables: 0/1 done 00:34:10.249 Creating journal (1024 blocks): done 00:34:10.249 Writing superblocks and filesystem accounting information: 0/1 done 00:34:10.249 00:34:10.249 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@142 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:34:10.249 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:34:10.249 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:34:10.249 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:34:10.249 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:34:10.249 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:34:10.249 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@327 -- # killprocess 91771 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@950 -- # '[' -z 91771 ']' 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@954 -- # kill -0 91771 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@955 -- # uname 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 91771 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:34:10.508 killing process with pid 91771 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@968 -- # echo 'killing process with pid 91771' 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@969 -- # kill 91771 00:34:10.508 20:33:05 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@974 -- # wait 91771 00:34:13.043 20:33:07 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@328 -- # trap - SIGINT SIGTERM EXIT 00:34:13.043 00:34:13.043 real 0m7.983s 00:34:13.043 user 0m10.968s 00:34:13.043 sys 0m1.624s 00:34:13.043 20:33:07 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:13.043 20:33:07 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:34:13.043 ************************************ 00:34:13.043 END TEST bdev_nbd 00:34:13.043 ************************************ 00:34:13.043 20:33:07 blockdev_raid5f -- bdev/blockdev.sh@764 -- # [[ y == y ]] 00:34:13.043 20:33:07 blockdev_raid5f -- bdev/blockdev.sh@765 -- # '[' raid5f = nvme ']' 00:34:13.043 20:33:07 blockdev_raid5f -- bdev/blockdev.sh@765 -- # '[' raid5f = gpt ']' 00:34:13.043 20:33:07 blockdev_raid5f -- bdev/blockdev.sh@769 -- # run_test bdev_fio fio_test_suite '' 00:34:13.043 20:33:07 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:34:13.043 20:33:07 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:34:13.043 20:33:07 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:34:13.043 ************************************ 00:34:13.043 START TEST bdev_fio 00:34:13.043 ************************************ 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1125 -- # fio_test_suite '' 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@332 -- # local env_context 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@336 -- # pushd /home/vagrant/spdk_repo/spdk/test/bdev 00:34:13.043 /home/vagrant/spdk_repo/spdk/test/bdev /home/vagrant/spdk_repo/spdk 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@337 -- # trap 'rm -f ./*.state; popd; exit 1' SIGINT SIGTERM EXIT 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@340 -- # echo '' 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@340 -- # sed s/--env-context=// 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@340 -- # env_context= 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@341 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio verify AIO '' 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1280 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1281 -- # local workload=verify 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1282 -- # local bdev_type=AIO 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1283 -- # local env_context= 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local fio_dir=/usr/src/fio 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1291 -- # '[' -z verify ']' 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -n '' ']' 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1301 -- # cat 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1313 -- # '[' verify == verify ']' 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1314 -- # cat 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1323 -- # '[' AIO == AIO ']' 00:34:13.043 20:33:07 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1324 -- # /usr/src/fio/fio --version 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1324 -- # [[ fio-3.35 == *\f\i\o\-\3* ]] 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1325 -- # echo serialize_overlap=1 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@342 -- # for b in "${bdevs_name[@]}" 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@343 -- # echo '[job_raid5f]' 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@344 -- # echo filename=raid5f 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@348 -- # local 'fio_params=--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json' 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@350 -- # run_test bdev_fio_rw_verify fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1101 -- # '[' 11 -le 1 ']' 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1107 -- # xtrace_disable 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:34:13.043 ************************************ 00:34:13.043 START TEST bdev_fio_rw_verify 00:34:13.043 ************************************ 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1125 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # fio_plugin /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # local sanitizers 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1340 -- # local plugin=/home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # shift 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # local asan_lib= 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # ldd /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:34:13.043 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # grep libasan 00:34:13.044 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:34:13.044 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # asan_lib=/usr/lib64/libasan.so.8 00:34:13.044 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1346 -- # [[ -n /usr/lib64/libasan.so.8 ]] 00:34:13.044 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1347 -- # break 00:34:13.044 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1352 -- # LD_PRELOAD='/usr/lib64/libasan.so.8 /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev' 00:34:13.044 20:33:08 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:34:13.303 job_raid5f: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:34:13.303 fio-3.35 00:34:13.303 Starting 1 thread 00:34:25.512 00:34:25.512 job_raid5f: (groupid=0, jobs=1): err= 0: pid=91996: Tue Oct 1 20:33:19 2024 00:34:25.512 read: IOPS=7884, BW=30.8MiB/s (32.3MB/s)(308MiB/10001msec) 00:34:25.512 slat (usec): min=22, max=121, avg=30.13, stdev= 7.66 00:34:25.512 clat (usec): min=14, max=570, avg=201.80, stdev=75.03 00:34:25.512 lat (usec): min=42, max=611, avg=231.93, stdev=76.36 00:34:25.512 clat percentiles (usec): 00:34:25.513 | 50.000th=[ 198], 99.000th=[ 375], 99.900th=[ 433], 99.990th=[ 498], 00:34:25.513 | 99.999th=[ 570] 00:34:25.513 write: IOPS=8256, BW=32.2MiB/s (33.8MB/s)(319MiB/9888msec); 0 zone resets 00:34:25.513 slat (usec): min=12, max=210, avg=25.89, stdev= 8.80 00:34:25.513 clat (usec): min=71, max=1328, avg=467.44, stdev=81.92 00:34:25.513 lat (usec): min=94, max=1539, avg=493.33, stdev=85.61 00:34:25.513 clat percentiles (usec): 00:34:25.513 | 50.000th=[ 461], 99.000th=[ 725], 99.900th=[ 881], 99.990th=[ 1074], 00:34:25.513 | 99.999th=[ 1336] 00:34:25.513 bw ( KiB/s): min=26952, max=36576, per=99.62%, avg=32900.21, stdev=2188.99, samples=19 00:34:25.513 iops : min= 6738, max= 9144, avg=8225.05, stdev=547.25, samples=19 00:34:25.513 lat (usec) : 20=0.01%, 100=4.73%, 250=29.82%, 500=51.43%, 750=13.67% 00:34:25.513 lat (usec) : 1000=0.33% 00:34:25.513 lat (msec) : 2=0.01% 00:34:25.513 cpu : usr=98.35%, sys=0.71%, ctx=21, majf=0, minf=6943 00:34:25.513 IO depths : 1=7.8%, 2=19.9%, 4=55.2%, 8=17.2%, 16=0.0%, 32=0.0%, >=64=0.0% 00:34:25.513 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:34:25.513 complete : 0=0.0%, 4=90.1%, 8=9.9%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:34:25.513 issued rwts: total=78856,81637,0,0 short=0,0,0,0 dropped=0,0,0,0 00:34:25.513 latency : target=0, window=0, percentile=100.00%, depth=8 00:34:25.513 00:34:25.513 Run status group 0 (all jobs): 00:34:25.513 READ: bw=30.8MiB/s (32.3MB/s), 30.8MiB/s-30.8MiB/s (32.3MB/s-32.3MB/s), io=308MiB (323MB), run=10001-10001msec 00:34:25.513 WRITE: bw=32.2MiB/s (33.8MB/s), 32.2MiB/s-32.2MiB/s (33.8MB/s-33.8MB/s), io=319MiB (334MB), run=9888-9888msec 00:34:26.466 ----------------------------------------------------- 00:34:26.466 Suppressions used: 00:34:26.466 count bytes template 00:34:26.466 1 7 /usr/src/fio/parse.c 00:34:26.466 329 31584 /usr/src/fio/iolog.c 00:34:26.466 1 8 libtcmalloc_minimal.so 00:34:26.466 1 904 libcrypto.so 00:34:26.466 ----------------------------------------------------- 00:34:26.466 00:34:26.466 00:34:26.466 real 0m13.652s 00:34:26.466 user 0m14.065s 00:34:26.466 sys 0m1.033s 00:34:26.466 20:33:21 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:26.466 20:33:21 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@10 -- # set +x 00:34:26.466 ************************************ 00:34:26.466 END TEST bdev_fio_rw_verify 00:34:26.466 ************************************ 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@351 -- # rm -f 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@352 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@355 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio trim '' '' 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1280 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1281 -- # local workload=trim 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1282 -- # local bdev_type= 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1283 -- # local env_context= 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local fio_dir=/usr/src/fio 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1291 -- # '[' -z trim ']' 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -n '' ']' 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1301 -- # cat 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1313 -- # '[' trim == verify ']' 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # '[' trim == trim ']' 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1329 -- # echo rw=trimwrite 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@356 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "9e952634-cec7-4ee5-b5dd-825034055a51"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "9e952634-cec7-4ee5-b5dd-825034055a51",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "9e952634-cec7-4ee5-b5dd-825034055a51",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "7a906a5c-da82-4570-8225-84806d3225ed",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "29123d65-a63b-44fb-9cc0-358d9ff3185e",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "bab1c060-9a38-421b-b614-13bb7f187036",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@356 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@356 -- # [[ -n '' ]] 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@362 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:34:26.733 /home/vagrant/spdk_repo/spdk 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@363 -- # popd 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@364 -- # trap - SIGINT SIGTERM EXIT 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@365 -- # return 0 00:34:26.733 00:34:26.733 real 0m13.882s 00:34:26.733 user 0m14.178s 00:34:26.733 sys 0m1.126s 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:26.733 20:33:21 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:34:26.733 ************************************ 00:34:26.733 END TEST bdev_fio 00:34:26.733 ************************************ 00:34:26.733 20:33:21 blockdev_raid5f -- bdev/blockdev.sh@776 -- # trap cleanup SIGINT SIGTERM EXIT 00:34:26.733 20:33:21 blockdev_raid5f -- bdev/blockdev.sh@778 -- # run_test bdev_verify /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:34:26.733 20:33:21 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 16 -le 1 ']' 00:34:26.733 20:33:21 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:34:26.733 20:33:21 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:34:26.733 ************************************ 00:34:26.733 START TEST bdev_verify 00:34:26.733 ************************************ 00:34:26.733 20:33:21 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:34:26.733 [2024-10-01 20:33:21.981762] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:34:26.733 [2024-10-01 20:33:21.981966] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92160 ] 00:34:26.993 [2024-10-01 20:33:22.160136] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:34:27.252 [2024-10-01 20:33:22.399175] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:34:27.252 [2024-10-01 20:33:22.399180] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:34:28.188 Running I/O for 5 seconds... 00:34:33.311 10497.00 IOPS, 41.00 MiB/s 9990.50 IOPS, 39.03 MiB/s 9714.00 IOPS, 37.95 MiB/s 9545.50 IOPS, 37.29 MiB/s 00:34:33.311 Latency(us) 00:34:33.312 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:34:33.312 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:34:33.312 Verification LBA range: start 0x0 length 0x2000 00:34:33.312 raid5f : 5.01 5009.71 19.57 0.00 0.00 38520.07 443.11 34078.72 00:34:33.312 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:34:33.312 Verification LBA range: start 0x2000 length 0x2000 00:34:33.312 raid5f : 5.01 4984.55 19.47 0.00 0.00 38763.67 299.75 35270.28 00:34:33.312 =================================================================================================================== 00:34:33.312 Total : 9994.26 39.04 0.00 0.00 38641.60 299.75 35270.28 00:34:35.218 00:34:35.218 real 0m8.548s 00:34:35.218 user 0m15.376s 00:34:35.218 sys 0m0.453s 00:34:35.218 20:33:30 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:35.218 ************************************ 00:34:35.218 END TEST bdev_verify 00:34:35.218 20:33:30 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@10 -- # set +x 00:34:35.218 ************************************ 00:34:35.218 20:33:30 blockdev_raid5f -- bdev/blockdev.sh@779 -- # run_test bdev_verify_big_io /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:34:35.218 20:33:30 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 16 -le 1 ']' 00:34:35.218 20:33:30 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:34:35.218 20:33:30 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:34:35.477 ************************************ 00:34:35.477 START TEST bdev_verify_big_io 00:34:35.477 ************************************ 00:34:35.477 20:33:30 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:34:35.477 [2024-10-01 20:33:30.597632] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:34:35.477 [2024-10-01 20:33:30.597895] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92264 ] 00:34:35.736 [2024-10-01 20:33:30.785787] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:34:35.996 [2024-10-01 20:33:31.083584] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:34:35.996 [2024-10-01 20:33:31.083591] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:34:36.932 Running I/O for 5 seconds... 00:34:42.175 505.00 IOPS, 31.56 MiB/s 538.00 IOPS, 33.62 MiB/s 592.00 IOPS, 37.00 MiB/s 634.00 IOPS, 39.62 MiB/s 647.00 IOPS, 40.44 MiB/s 00:34:42.175 Latency(us) 00:34:42.175 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:34:42.175 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:34:42.175 Verification LBA range: start 0x0 length 0x200 00:34:42.175 raid5f : 5.30 323.50 20.22 0.00 0.00 9668482.33 359.33 425149.91 00:34:42.175 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:34:42.175 Verification LBA range: start 0x200 length 0x200 00:34:42.175 raid5f : 5.34 332.78 20.80 0.00 0.00 9426503.59 288.58 419430.40 00:34:42.175 =================================================================================================================== 00:34:42.175 Total : 656.29 41.02 0.00 0.00 9545308.28 288.58 425149.91 00:34:44.708 00:34:44.708 real 0m9.252s 00:34:44.708 user 0m16.632s 00:34:44.708 sys 0m0.512s 00:34:44.708 20:33:39 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:44.708 ************************************ 00:34:44.708 END TEST bdev_verify_big_io 00:34:44.708 ************************************ 00:34:44.708 20:33:39 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@10 -- # set +x 00:34:44.708 20:33:39 blockdev_raid5f -- bdev/blockdev.sh@780 -- # run_test bdev_write_zeroes /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:34:44.708 20:33:39 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 13 -le 1 ']' 00:34:44.708 20:33:39 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:34:44.708 20:33:39 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:34:44.708 ************************************ 00:34:44.708 START TEST bdev_write_zeroes 00:34:44.708 ************************************ 00:34:44.708 20:33:39 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:34:44.708 [2024-10-01 20:33:39.888920] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:34:44.708 [2024-10-01 20:33:39.889124] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92382 ] 00:34:44.967 [2024-10-01 20:33:40.070825] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:45.225 [2024-10-01 20:33:40.349593] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:34:46.160 Running I/O for 1 seconds... 00:34:47.094 16695.00 IOPS, 65.21 MiB/s 00:34:47.094 Latency(us) 00:34:47.094 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:34:47.094 Job: raid5f (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:34:47.094 raid5f : 1.01 16673.41 65.13 0.00 0.00 7641.08 2621.44 10247.45 00:34:47.094 =================================================================================================================== 00:34:47.094 Total : 16673.41 65.13 0.00 0.00 7641.08 2621.44 10247.45 00:34:49.634 00:34:49.634 real 0m4.896s 00:34:49.634 user 0m4.287s 00:34:49.634 sys 0m0.463s 00:34:49.634 20:33:44 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:49.634 20:33:44 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@10 -- # set +x 00:34:49.634 ************************************ 00:34:49.634 END TEST bdev_write_zeroes 00:34:49.634 ************************************ 00:34:49.634 20:33:44 blockdev_raid5f -- bdev/blockdev.sh@783 -- # run_test bdev_json_nonenclosed /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:34:49.634 20:33:44 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 13 -le 1 ']' 00:34:49.634 20:33:44 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:34:49.634 20:33:44 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:34:49.634 ************************************ 00:34:49.634 START TEST bdev_json_nonenclosed 00:34:49.634 ************************************ 00:34:49.634 20:33:44 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:34:49.634 [2024-10-01 20:33:44.842628] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:34:49.634 [2024-10-01 20:33:44.842869] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92446 ] 00:34:49.893 [2024-10-01 20:33:45.025658] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:50.154 [2024-10-01 20:33:45.344552] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:34:50.154 [2024-10-01 20:33:45.344727] json_config.c: 608:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: not enclosed in {}. 00:34:50.154 [2024-10-01 20:33:45.344819] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:34:50.154 [2024-10-01 20:33:45.344852] app.c:1061:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:34:50.731 00:34:50.731 real 0m1.110s 00:34:50.731 user 0m0.806s 00:34:50.731 sys 0m0.193s 00:34:50.731 20:33:45 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:50.731 20:33:45 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@10 -- # set +x 00:34:50.731 ************************************ 00:34:50.731 END TEST bdev_json_nonenclosed 00:34:50.731 ************************************ 00:34:50.731 20:33:45 blockdev_raid5f -- bdev/blockdev.sh@786 -- # run_test bdev_json_nonarray /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:34:50.731 20:33:45 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 13 -le 1 ']' 00:34:50.731 20:33:45 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:34:50.731 20:33:45 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:34:50.731 ************************************ 00:34:50.731 START TEST bdev_json_nonarray 00:34:50.731 ************************************ 00:34:50.731 20:33:45 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:34:50.990 [2024-10-01 20:33:46.004641] Starting SPDK v25.01-pre git sha1 1b1c3081e / DPDK 24.03.0 initialization... 00:34:50.990 [2024-10-01 20:33:46.004896] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92482 ] 00:34:50.990 [2024-10-01 20:33:46.189420] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:51.248 [2024-10-01 20:33:46.482965] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:34:51.248 [2024-10-01 20:33:46.483169] json_config.c: 614:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: 'subsystems' should be an array. 00:34:51.248 [2024-10-01 20:33:46.483227] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:34:51.248 [2024-10-01 20:33:46.483258] app.c:1061:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:34:51.816 00:34:51.816 real 0m1.058s 00:34:51.816 user 0m0.772s 00:34:51.817 sys 0m0.178s 00:34:51.817 20:33:46 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:51.817 20:33:46 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@10 -- # set +x 00:34:51.817 ************************************ 00:34:51.817 END TEST bdev_json_nonarray 00:34:51.817 ************************************ 00:34:51.817 20:33:46 blockdev_raid5f -- bdev/blockdev.sh@788 -- # [[ raid5f == bdev ]] 00:34:51.817 20:33:46 blockdev_raid5f -- bdev/blockdev.sh@795 -- # [[ raid5f == gpt ]] 00:34:51.817 20:33:46 blockdev_raid5f -- bdev/blockdev.sh@799 -- # [[ raid5f == crypto_sw ]] 00:34:51.817 20:33:46 blockdev_raid5f -- bdev/blockdev.sh@811 -- # trap - SIGINT SIGTERM EXIT 00:34:51.817 20:33:46 blockdev_raid5f -- bdev/blockdev.sh@812 -- # cleanup 00:34:51.817 20:33:46 blockdev_raid5f -- bdev/blockdev.sh@23 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/aiofile 00:34:51.817 20:33:46 blockdev_raid5f -- bdev/blockdev.sh@24 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:34:51.817 20:33:46 blockdev_raid5f -- bdev/blockdev.sh@26 -- # [[ raid5f == rbd ]] 00:34:51.817 20:33:46 blockdev_raid5f -- bdev/blockdev.sh@30 -- # [[ raid5f == daos ]] 00:34:51.817 20:33:46 blockdev_raid5f -- bdev/blockdev.sh@34 -- # [[ raid5f = \g\p\t ]] 00:34:51.817 20:33:46 blockdev_raid5f -- bdev/blockdev.sh@40 -- # [[ raid5f == xnvme ]] 00:34:51.817 00:34:51.817 real 1m0.689s 00:34:51.817 user 1m21.691s 00:34:51.817 sys 0m6.811s 00:34:51.817 ************************************ 00:34:51.817 END TEST blockdev_raid5f 00:34:51.817 ************************************ 00:34:51.817 20:33:46 blockdev_raid5f -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:51.817 20:33:46 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:34:51.817 20:33:47 -- spdk/autotest.sh@194 -- # uname -s 00:34:51.817 20:33:47 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:34:51.817 20:33:47 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:34:51.817 20:33:47 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:34:51.817 20:33:47 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:34:51.817 20:33:47 -- spdk/autotest.sh@252 -- # '[' 0 -eq 1 ']' 00:34:51.817 20:33:47 -- spdk/autotest.sh@256 -- # timing_exit lib 00:34:51.817 20:33:47 -- common/autotest_common.sh@730 -- # xtrace_disable 00:34:51.817 20:33:47 -- common/autotest_common.sh@10 -- # set +x 00:34:52.076 20:33:47 -- spdk/autotest.sh@258 -- # '[' 0 -eq 1 ']' 00:34:52.076 20:33:47 -- spdk/autotest.sh@263 -- # '[' 0 -eq 1 ']' 00:34:52.076 20:33:47 -- spdk/autotest.sh@272 -- # '[' 0 -eq 1 ']' 00:34:52.076 20:33:47 -- spdk/autotest.sh@307 -- # '[' 0 -eq 1 ']' 00:34:52.076 20:33:47 -- spdk/autotest.sh@311 -- # '[' 0 -eq 1 ']' 00:34:52.076 20:33:47 -- spdk/autotest.sh@315 -- # '[' 0 -eq 1 ']' 00:34:52.076 20:33:47 -- spdk/autotest.sh@320 -- # '[' 0 -eq 1 ']' 00:34:52.076 20:33:47 -- spdk/autotest.sh@329 -- # '[' 0 -eq 1 ']' 00:34:52.076 20:33:47 -- spdk/autotest.sh@334 -- # '[' 0 -eq 1 ']' 00:34:52.076 20:33:47 -- spdk/autotest.sh@338 -- # '[' 0 -eq 1 ']' 00:34:52.076 20:33:47 -- spdk/autotest.sh@342 -- # '[' 0 -eq 1 ']' 00:34:52.076 20:33:47 -- spdk/autotest.sh@346 -- # '[' 0 -eq 1 ']' 00:34:52.076 20:33:47 -- spdk/autotest.sh@351 -- # '[' 0 -eq 1 ']' 00:34:52.076 20:33:47 -- spdk/autotest.sh@355 -- # '[' 0 -eq 1 ']' 00:34:52.076 20:33:47 -- spdk/autotest.sh@362 -- # [[ 0 -eq 1 ]] 00:34:52.077 20:33:47 -- spdk/autotest.sh@366 -- # [[ 0 -eq 1 ]] 00:34:52.077 20:33:47 -- spdk/autotest.sh@370 -- # [[ 0 -eq 1 ]] 00:34:52.077 20:33:47 -- spdk/autotest.sh@374 -- # [[ '' -eq 1 ]] 00:34:52.077 20:33:47 -- spdk/autotest.sh@381 -- # trap - SIGINT SIGTERM EXIT 00:34:52.077 20:33:47 -- spdk/autotest.sh@383 -- # timing_enter post_cleanup 00:34:52.077 20:33:47 -- common/autotest_common.sh@724 -- # xtrace_disable 00:34:52.077 20:33:47 -- common/autotest_common.sh@10 -- # set +x 00:34:52.077 20:33:47 -- spdk/autotest.sh@384 -- # autotest_cleanup 00:34:52.077 20:33:47 -- common/autotest_common.sh@1392 -- # local autotest_es=0 00:34:52.077 20:33:47 -- common/autotest_common.sh@1393 -- # xtrace_disable 00:34:52.077 20:33:47 -- common/autotest_common.sh@10 -- # set +x 00:34:53.982 INFO: APP EXITING 00:34:53.982 INFO: killing all VMs 00:34:53.982 INFO: killing vhost app 00:34:53.982 INFO: EXIT DONE 00:34:53.982 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:34:53.982 Waiting for block devices as requested 00:34:53.982 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:34:54.241 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:34:54.810 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:34:55.069 Cleaning 00:34:55.069 Removing: /var/run/dpdk/spdk0/config 00:34:55.069 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:34:55.069 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:34:55.069 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:34:55.069 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:34:55.069 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:34:55.069 Removing: /var/run/dpdk/spdk0/hugepage_info 00:34:55.070 Removing: /dev/shm/spdk_tgt_trace.pid56831 00:34:55.070 Removing: /var/run/dpdk/spdk0 00:34:55.070 Removing: /var/run/dpdk/spdk_pid56596 00:34:55.070 Removing: /var/run/dpdk/spdk_pid56831 00:34:55.070 Removing: /var/run/dpdk/spdk_pid57072 00:34:55.070 Removing: /var/run/dpdk/spdk_pid57186 00:34:55.070 Removing: /var/run/dpdk/spdk_pid57254 00:34:55.070 Removing: /var/run/dpdk/spdk_pid57393 00:34:55.070 Removing: /var/run/dpdk/spdk_pid57417 00:34:55.070 Removing: /var/run/dpdk/spdk_pid57638 00:34:55.070 Removing: /var/run/dpdk/spdk_pid57768 00:34:55.070 Removing: /var/run/dpdk/spdk_pid57886 00:34:55.070 Removing: /var/run/dpdk/spdk_pid58019 00:34:55.070 Removing: /var/run/dpdk/spdk_pid58138 00:34:55.070 Removing: /var/run/dpdk/spdk_pid58183 00:34:55.070 Removing: /var/run/dpdk/spdk_pid58225 00:34:55.070 Removing: /var/run/dpdk/spdk_pid58301 00:34:55.070 Removing: /var/run/dpdk/spdk_pid58424 00:34:55.070 Removing: /var/run/dpdk/spdk_pid58920 00:34:55.070 Removing: /var/run/dpdk/spdk_pid59007 00:34:55.070 Removing: /var/run/dpdk/spdk_pid59093 00:34:55.070 Removing: /var/run/dpdk/spdk_pid59120 00:34:55.070 Removing: /var/run/dpdk/spdk_pid59301 00:34:55.070 Removing: /var/run/dpdk/spdk_pid59323 00:34:55.070 Removing: /var/run/dpdk/spdk_pid59505 00:34:55.070 Removing: /var/run/dpdk/spdk_pid59532 00:34:55.070 Removing: /var/run/dpdk/spdk_pid59607 00:34:55.070 Removing: /var/run/dpdk/spdk_pid59636 00:34:55.070 Removing: /var/run/dpdk/spdk_pid59711 00:34:55.070 Removing: /var/run/dpdk/spdk_pid59735 00:34:55.070 Removing: /var/run/dpdk/spdk_pid59952 00:34:55.070 Removing: /var/run/dpdk/spdk_pid59994 00:34:55.070 Removing: /var/run/dpdk/spdk_pid60083 00:34:55.070 Removing: /var/run/dpdk/spdk_pid61542 00:34:55.070 Removing: /var/run/dpdk/spdk_pid61770 00:34:55.070 Removing: /var/run/dpdk/spdk_pid61921 00:34:55.070 Removing: /var/run/dpdk/spdk_pid62603 00:34:55.070 Removing: /var/run/dpdk/spdk_pid62826 00:34:55.070 Removing: /var/run/dpdk/spdk_pid62977 00:34:55.070 Removing: /var/run/dpdk/spdk_pid63659 00:34:55.070 Removing: /var/run/dpdk/spdk_pid64001 00:34:55.070 Removing: /var/run/dpdk/spdk_pid64158 00:34:55.070 Removing: /var/run/dpdk/spdk_pid65604 00:34:55.070 Removing: /var/run/dpdk/spdk_pid65868 00:34:55.070 Removing: /var/run/dpdk/spdk_pid66025 00:34:55.070 Removing: /var/run/dpdk/spdk_pid67460 00:34:55.070 Removing: /var/run/dpdk/spdk_pid67724 00:34:55.070 Removing: /var/run/dpdk/spdk_pid67885 00:34:55.070 Removing: /var/run/dpdk/spdk_pid69326 00:34:55.070 Removing: /var/run/dpdk/spdk_pid69789 00:34:55.070 Removing: /var/run/dpdk/spdk_pid69940 00:34:55.070 Removing: /var/run/dpdk/spdk_pid71470 00:34:55.070 Removing: /var/run/dpdk/spdk_pid71748 00:34:55.070 Removing: /var/run/dpdk/spdk_pid71903 00:34:55.070 Removing: /var/run/dpdk/spdk_pid73440 00:34:55.070 Removing: /var/run/dpdk/spdk_pid73716 00:34:55.070 Removing: /var/run/dpdk/spdk_pid73874 00:34:55.070 Removing: /var/run/dpdk/spdk_pid75408 00:34:55.070 Removing: /var/run/dpdk/spdk_pid75911 00:34:55.070 Removing: /var/run/dpdk/spdk_pid76069 00:34:55.070 Removing: /var/run/dpdk/spdk_pid76217 00:34:55.070 Removing: /var/run/dpdk/spdk_pid76682 00:34:55.070 Removing: /var/run/dpdk/spdk_pid77451 00:34:55.070 Removing: /var/run/dpdk/spdk_pid77867 00:34:55.070 Removing: /var/run/dpdk/spdk_pid78599 00:34:55.070 Removing: /var/run/dpdk/spdk_pid79094 00:34:55.330 Removing: /var/run/dpdk/spdk_pid79904 00:34:55.330 Removing: /var/run/dpdk/spdk_pid80349 00:34:55.330 Removing: /var/run/dpdk/spdk_pid82375 00:34:55.330 Removing: /var/run/dpdk/spdk_pid82836 00:34:55.330 Removing: /var/run/dpdk/spdk_pid83296 00:34:55.330 Removing: /var/run/dpdk/spdk_pid85460 00:34:55.330 Removing: /var/run/dpdk/spdk_pid85968 00:34:55.330 Removing: /var/run/dpdk/spdk_pid86488 00:34:55.330 Removing: /var/run/dpdk/spdk_pid87597 00:34:55.330 Removing: /var/run/dpdk/spdk_pid87931 00:34:55.330 Removing: /var/run/dpdk/spdk_pid88915 00:34:55.330 Removing: /var/run/dpdk/spdk_pid89253 00:34:55.330 Removing: /var/run/dpdk/spdk_pid90243 00:34:55.330 Removing: /var/run/dpdk/spdk_pid90583 00:34:55.330 Removing: /var/run/dpdk/spdk_pid91271 00:34:55.330 Removing: /var/run/dpdk/spdk_pid91563 00:34:55.330 Removing: /var/run/dpdk/spdk_pid91641 00:34:55.330 Removing: /var/run/dpdk/spdk_pid91699 00:34:55.330 Removing: /var/run/dpdk/spdk_pid91977 00:34:55.330 Removing: /var/run/dpdk/spdk_pid92160 00:34:55.330 Removing: /var/run/dpdk/spdk_pid92264 00:34:55.330 Removing: /var/run/dpdk/spdk_pid92382 00:34:55.330 Removing: /var/run/dpdk/spdk_pid92446 00:34:55.330 Removing: /var/run/dpdk/spdk_pid92482 00:34:55.330 Clean 00:34:55.330 20:33:50 -- common/autotest_common.sh@1451 -- # return 0 00:34:55.330 20:33:50 -- spdk/autotest.sh@385 -- # timing_exit post_cleanup 00:34:55.330 20:33:50 -- common/autotest_common.sh@730 -- # xtrace_disable 00:34:55.330 20:33:50 -- common/autotest_common.sh@10 -- # set +x 00:34:55.330 20:33:50 -- spdk/autotest.sh@387 -- # timing_exit autotest 00:34:55.330 20:33:50 -- common/autotest_common.sh@730 -- # xtrace_disable 00:34:55.330 20:33:50 -- common/autotest_common.sh@10 -- # set +x 00:34:55.330 20:33:50 -- spdk/autotest.sh@388 -- # chmod a+r /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:34:55.330 20:33:50 -- spdk/autotest.sh@390 -- # [[ -f /home/vagrant/spdk_repo/spdk/../output/udev.log ]] 00:34:55.330 20:33:50 -- spdk/autotest.sh@390 -- # rm -f /home/vagrant/spdk_repo/spdk/../output/udev.log 00:34:55.330 20:33:50 -- spdk/autotest.sh@392 -- # [[ y == y ]] 00:34:55.330 20:33:50 -- spdk/autotest.sh@394 -- # hostname 00:34:55.330 20:33:50 -- spdk/autotest.sh@394 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -d /home/vagrant/spdk_repo/spdk -t fedora39-cloud-1721788873-2326 -o /home/vagrant/spdk_repo/spdk/../output/cov_test.info 00:34:55.589 geninfo: WARNING: invalid characters removed from testname! 00:35:22.190 20:34:17 -- spdk/autotest.sh@395 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -a /home/vagrant/spdk_repo/spdk/../output/cov_base.info -a /home/vagrant/spdk_repo/spdk/../output/cov_test.info -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:35:27.465 20:34:21 -- spdk/autotest.sh@396 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/dpdk/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:35:30.002 20:34:24 -- spdk/autotest.sh@400 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info --ignore-errors unused,unused '/usr/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:35:33.294 20:34:27 -- spdk/autotest.sh@401 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/examples/vmd/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:35:35.832 20:34:30 -- spdk/autotest.sh@402 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:35:39.123 20:34:33 -- spdk/autotest.sh@403 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:35:42.412 20:34:37 -- spdk/autotest.sh@404 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:35:42.412 20:34:37 -- common/autotest_common.sh@1680 -- $ [[ y == y ]] 00:35:42.412 20:34:37 -- common/autotest_common.sh@1681 -- $ lcov --version 00:35:42.412 20:34:37 -- common/autotest_common.sh@1681 -- $ awk '{print $NF}' 00:35:42.412 20:34:37 -- common/autotest_common.sh@1681 -- $ lt 1.15 2 00:35:42.412 20:34:37 -- scripts/common.sh@373 -- $ cmp_versions 1.15 '<' 2 00:35:42.412 20:34:37 -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:35:42.412 20:34:37 -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:35:42.412 20:34:37 -- scripts/common.sh@336 -- $ IFS=.-: 00:35:42.412 20:34:37 -- scripts/common.sh@336 -- $ read -ra ver1 00:35:42.412 20:34:37 -- scripts/common.sh@337 -- $ IFS=.-: 00:35:42.412 20:34:37 -- scripts/common.sh@337 -- $ read -ra ver2 00:35:42.412 20:34:37 -- scripts/common.sh@338 -- $ local 'op=<' 00:35:42.412 20:34:37 -- scripts/common.sh@340 -- $ ver1_l=2 00:35:42.412 20:34:37 -- scripts/common.sh@341 -- $ ver2_l=1 00:35:42.412 20:34:37 -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:35:42.412 20:34:37 -- scripts/common.sh@344 -- $ case "$op" in 00:35:42.412 20:34:37 -- scripts/common.sh@345 -- $ : 1 00:35:42.412 20:34:37 -- scripts/common.sh@364 -- $ (( v = 0 )) 00:35:42.412 20:34:37 -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:35:42.412 20:34:37 -- scripts/common.sh@365 -- $ decimal 1 00:35:42.412 20:34:37 -- scripts/common.sh@353 -- $ local d=1 00:35:42.412 20:34:37 -- scripts/common.sh@354 -- $ [[ 1 =~ ^[0-9]+$ ]] 00:35:42.412 20:34:37 -- scripts/common.sh@355 -- $ echo 1 00:35:42.412 20:34:37 -- scripts/common.sh@365 -- $ ver1[v]=1 00:35:42.412 20:34:37 -- scripts/common.sh@366 -- $ decimal 2 00:35:42.412 20:34:37 -- scripts/common.sh@353 -- $ local d=2 00:35:42.412 20:34:37 -- scripts/common.sh@354 -- $ [[ 2 =~ ^[0-9]+$ ]] 00:35:42.412 20:34:37 -- scripts/common.sh@355 -- $ echo 2 00:35:42.412 20:34:37 -- scripts/common.sh@366 -- $ ver2[v]=2 00:35:42.412 20:34:37 -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:35:42.412 20:34:37 -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:35:42.412 20:34:37 -- scripts/common.sh@368 -- $ return 0 00:35:42.412 20:34:37 -- common/autotest_common.sh@1682 -- $ lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:35:42.412 20:34:37 -- common/autotest_common.sh@1694 -- $ export 'LCOV_OPTS= 00:35:42.412 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:42.412 --rc genhtml_branch_coverage=1 00:35:42.412 --rc genhtml_function_coverage=1 00:35:42.412 --rc genhtml_legend=1 00:35:42.412 --rc geninfo_all_blocks=1 00:35:42.412 --rc geninfo_unexecuted_blocks=1 00:35:42.412 00:35:42.412 ' 00:35:42.412 20:34:37 -- common/autotest_common.sh@1694 -- $ LCOV_OPTS=' 00:35:42.412 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:42.412 --rc genhtml_branch_coverage=1 00:35:42.412 --rc genhtml_function_coverage=1 00:35:42.412 --rc genhtml_legend=1 00:35:42.412 --rc geninfo_all_blocks=1 00:35:42.412 --rc geninfo_unexecuted_blocks=1 00:35:42.412 00:35:42.412 ' 00:35:42.412 20:34:37 -- common/autotest_common.sh@1695 -- $ export 'LCOV=lcov 00:35:42.412 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:42.412 --rc genhtml_branch_coverage=1 00:35:42.412 --rc genhtml_function_coverage=1 00:35:42.412 --rc genhtml_legend=1 00:35:42.412 --rc geninfo_all_blocks=1 00:35:42.412 --rc geninfo_unexecuted_blocks=1 00:35:42.412 00:35:42.412 ' 00:35:42.412 20:34:37 -- common/autotest_common.sh@1695 -- $ LCOV='lcov 00:35:42.412 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:42.412 --rc genhtml_branch_coverage=1 00:35:42.412 --rc genhtml_function_coverage=1 00:35:42.412 --rc genhtml_legend=1 00:35:42.412 --rc geninfo_all_blocks=1 00:35:42.412 --rc geninfo_unexecuted_blocks=1 00:35:42.412 00:35:42.412 ' 00:35:42.412 20:34:37 -- common/autobuild_common.sh@15 -- $ source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:35:42.412 20:34:37 -- scripts/common.sh@15 -- $ shopt -s extglob 00:35:42.412 20:34:37 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:35:42.412 20:34:37 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:35:42.412 20:34:37 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:35:42.412 20:34:37 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:42.412 20:34:37 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:42.413 20:34:37 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:42.413 20:34:37 -- paths/export.sh@5 -- $ export PATH 00:35:42.413 20:34:37 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:42.413 20:34:37 -- common/autobuild_common.sh@478 -- $ out=/home/vagrant/spdk_repo/spdk/../output 00:35:42.413 20:34:37 -- common/autobuild_common.sh@479 -- $ date +%s 00:35:42.413 20:34:37 -- common/autobuild_common.sh@479 -- $ mktemp -dt spdk_1727814877.XXXXXX 00:35:42.413 20:34:37 -- common/autobuild_common.sh@479 -- $ SPDK_WORKSPACE=/tmp/spdk_1727814877.pZkFBR 00:35:42.413 20:34:37 -- common/autobuild_common.sh@481 -- $ [[ -n '' ]] 00:35:42.413 20:34:37 -- common/autobuild_common.sh@485 -- $ '[' -n '' ']' 00:35:42.413 20:34:37 -- common/autobuild_common.sh@488 -- $ scanbuild_exclude='--exclude /home/vagrant/spdk_repo/spdk/dpdk/' 00:35:42.413 20:34:37 -- common/autobuild_common.sh@492 -- $ scanbuild_exclude+=' --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp' 00:35:42.413 20:34:37 -- common/autobuild_common.sh@494 -- $ scanbuild='scan-build -o /home/vagrant/spdk_repo/spdk/../output/scan-build-tmp --exclude /home/vagrant/spdk_repo/spdk/dpdk/ --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp --status-bugs' 00:35:42.413 20:34:37 -- common/autobuild_common.sh@495 -- $ get_config_params 00:35:42.413 20:34:37 -- common/autotest_common.sh@407 -- $ xtrace_disable 00:35:42.413 20:34:37 -- common/autotest_common.sh@10 -- $ set +x 00:35:42.413 20:34:37 -- common/autobuild_common.sh@495 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f' 00:35:42.413 20:34:37 -- common/autobuild_common.sh@497 -- $ start_monitor_resources 00:35:42.413 20:34:37 -- pm/common@17 -- $ local monitor 00:35:42.413 20:34:37 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:35:42.413 20:34:37 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:35:42.413 20:34:37 -- pm/common@25 -- $ sleep 1 00:35:42.413 20:34:37 -- pm/common@21 -- $ date +%s 00:35:42.413 20:34:37 -- pm/common@21 -- $ date +%s 00:35:42.413 20:34:37 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autopackage.sh.1727814877 00:35:42.413 20:34:37 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autopackage.sh.1727814877 00:35:42.413 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autopackage.sh.1727814877_collect-cpu-load.pm.log 00:35:42.413 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autopackage.sh.1727814877_collect-vmstat.pm.log 00:35:43.350 20:34:38 -- common/autobuild_common.sh@498 -- $ trap stop_monitor_resources EXIT 00:35:43.351 20:34:38 -- spdk/autopackage.sh@10 -- $ [[ 0 -eq 1 ]] 00:35:43.351 20:34:38 -- spdk/autopackage.sh@14 -- $ timing_finish 00:35:43.351 20:34:38 -- common/autotest_common.sh@736 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:35:43.351 20:34:38 -- common/autotest_common.sh@737 -- $ [[ -x /usr/local/FlameGraph/flamegraph.pl ]] 00:35:43.351 20:34:38 -- common/autotest_common.sh@740 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:35:43.351 20:34:38 -- spdk/autopackage.sh@1 -- $ stop_monitor_resources 00:35:43.351 20:34:38 -- pm/common@29 -- $ signal_monitor_resources TERM 00:35:43.351 20:34:38 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:35:43.351 20:34:38 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:35:43.351 20:34:38 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-cpu-load.pid ]] 00:35:43.351 20:34:38 -- pm/common@44 -- $ pid=93974 00:35:43.351 20:34:38 -- pm/common@50 -- $ kill -TERM 93974 00:35:43.351 20:34:38 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:35:43.351 20:34:38 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-vmstat.pid ]] 00:35:43.351 20:34:38 -- pm/common@44 -- $ pid=93976 00:35:43.351 20:34:38 -- pm/common@50 -- $ kill -TERM 93976 00:35:43.351 + [[ -n 5207 ]] 00:35:43.351 + sudo kill 5207 00:35:43.360 [Pipeline] } 00:35:43.377 [Pipeline] // timeout 00:35:43.382 [Pipeline] } 00:35:43.398 [Pipeline] // stage 00:35:43.403 [Pipeline] } 00:35:43.417 [Pipeline] // catchError 00:35:43.427 [Pipeline] stage 00:35:43.429 [Pipeline] { (Stop VM) 00:35:43.442 [Pipeline] sh 00:35:43.723 + vagrant halt 00:35:47.012 ==> default: Halting domain... 00:35:52.302 [Pipeline] sh 00:35:52.581 + vagrant destroy -f 00:35:55.867 ==> default: Removing domain... 00:35:55.882 [Pipeline] sh 00:35:56.173 + mv output /var/jenkins/workspace/raid-vg-autotest_3/output 00:35:56.183 [Pipeline] } 00:35:56.203 [Pipeline] // stage 00:35:56.211 [Pipeline] } 00:35:56.226 [Pipeline] // dir 00:35:56.233 [Pipeline] } 00:35:56.250 [Pipeline] // wrap 00:35:56.257 [Pipeline] } 00:35:56.270 [Pipeline] // catchError 00:35:56.279 [Pipeline] stage 00:35:56.281 [Pipeline] { (Epilogue) 00:35:56.294 [Pipeline] sh 00:35:56.576 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:36:01.859 [Pipeline] catchError 00:36:01.862 [Pipeline] { 00:36:01.875 [Pipeline] sh 00:36:02.154 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:36:02.154 Artifacts sizes are good 00:36:02.162 [Pipeline] } 00:36:02.177 [Pipeline] // catchError 00:36:02.189 [Pipeline] archiveArtifacts 00:36:02.196 Archiving artifacts 00:36:02.306 [Pipeline] cleanWs 00:36:02.318 [WS-CLEANUP] Deleting project workspace... 00:36:02.318 [WS-CLEANUP] Deferred wipeout is used... 00:36:02.325 [WS-CLEANUP] done 00:36:02.327 [Pipeline] } 00:36:02.343 [Pipeline] // stage 00:36:02.348 [Pipeline] } 00:36:02.363 [Pipeline] // node 00:36:02.373 [Pipeline] End of Pipeline 00:36:02.431 Finished: SUCCESS